I am creating my own custom shapes for some of my enterprise architecture works and thought of taking the stencil route, the question i have is if i am using custom shapes from my stencils in my visio diagrams, and later on if i go and update those shapes in stencils, does those change reflect back to all diagrams? Is this possible?
When you modify the masters in a document all instances of those masters in the same document will be updated. Shapes you created in other documents will not be updated because when you drag-and-drop from a stencil to another document the master is copied into the document.
If you are modifying masters using the automation API make sure you call Master.Open and Master.Close. If you don’t changes to groups and shape counts won’t propagate to instance shapes.
Related
I'm looking to mask all of the detail in Tableau maps other than the state I'm filtered on. The data layer can seemingly only be applied to all or nothing. I'm unable to filter or limit to the dataset to which I'm working. Seems simple, yet 30min of google search can't find answer. Thank you!
I don’t believe that the map server can alter what it displays as a data layer based on the filters you choose in Tableau. The map server is a separate program that returns image tiles upon request. The map server doesn’t know about your filters on the client side - or even that the client app is called Tableau. It just knows that some client application asked for an image representing a map within a particular coordinate bounding box, with whatever style and data layer options that the client application passed.
That doesn’t mean you can’t accomplish your goal, just that you have to find a way to accomplish it on the Tableau side instead the map server side — unless you want to develop your own map server with that functionality, which would be an effort even with a tool like Mapbox.
You could
Decide what “data layers” the map server is providing that you want to retain and at what level of detail, say Avg(Income) per County
Go get that data, possibly from the census bureau
Uncheck the box for your data layers in the Map Layers pane
Use the data you downloaded above to color the map sections as desired
Add other marks to represent additional info as needed using a dual-axis and possibly data blending
Then voila, your data layers will obey your filter. You just had to move where that info was drawn. Of course, you had to find the necessary data instead of relying on the whoever built the map server to do that, but that isn’t too tough.
To handle the Tableau side of things, especially assuming you want to combine the layer we’ve been discussing with other data on your map, you will need to learn about a few other Tableau concepts. So read up on dual-axis maps, level of detail and possibly data blending.
Try this-Right click on the filter state and click show filter.Again right click on filter select "Multiple values(drop down) and then click the Customize and show apply button.Then choose the filter values and apply filter.enter image description here
I'm new for unity.. In my project I have a function that indicate Interior designing using Augment reality.. The thing is the designer can be able to upload their designed interior design in to our app, for this I need some help, When I refer the internet I got the information about 'online Directory' (For store the asset in to that, then he can upload that), I don't know whether it will be useful or not, Can any one please suggest a solution for this scenario?
First of all it is important to know that the user can not send a 3D model (.FBX) to the app. You can only use existing models within Unity 3D.
If you want the user to download new 3D models, you can use AssetBundles.
If you want the user to create a decoration and share it with friends, you should create a JSON / XML with all objects and positions and save it to your database. When another user requests this file, it will be downloaded and the objects will be dynamically created in a scene for the user to admire all 3D objects.
In cq we can create live copies by either blueprints by opting "New Site..." or directly trough content nodes by opting for "New Live Copy...".
In both the cases inheritance is maintained and roll-out works too in same ways. So what is the advantage of using one over other.
Any views?
Live Copies
Live copies can be created for just a simple page or a tree of pages and might the page and it's subpages depending on the rollout configuration. A live copy can be linked to a rollout config or will use the system's default one.
There is no formal requirement on the source page's structure.
A live copy might reference a blueprint, while it can only reference to a single blueprint.
Blueprints
Blueprints target the rollout of complete multilingual website projects and are a tool to control multiple rollout configs and live copies.
A blueprint requires a certain structure for the source site:
- One root level page
- The root's immediate childs define the language branches of the site
- each language contains one or more child pages.
Blueprints allow you to control multiple live copies and centrally consistent rollout configs for the blueprint's live copies.
A blueprint rollout will push modifications to all it's live copies.
Usage scenarios of blueprints
Inheritance and rollout work the same way. Just because blueprint make use of live copies.
But blueprints help you to organize your rollout scenarios for large multilingual sites. Just imagine a corporate website that provides a two or even three digit number of locales which that need to be translated and kept in sync.
In such a scenario you will likely end up with a hardly understandable and maintainable number of live copy and rollout configurations.
Depending on a blueprint to e.g. standardize the rollout of a new language/market/locale provides you higher degree of governance over your process as the complete process centrally manageable through the blueprint template.
But as long as you do not have such a scenario you might be fine without having the complete blueprint overhead.
A Livecopy is defined in the target page node with a cq:LiveSyncConfig node. It basically defines: I am a livecopy of source (blueprint) page X, and the following rollout configs apply.
A Blueprint is defined in the source page node with a cq:BlueprintSyncConfig node, and this defines a target.
Essentially both achieve the same in the end, but I think there are a few differences: the first option can be used to create a 1:n relationship, whereas the second option is 1:1
Also, if page nodes are copy-pasted in AEM, then relationships are copied with them (not quite sure in which way exactly, you would have to try for both scenarios). Also when pages are deleted in a tree in the first scenario, AEM will add a cq:excludedPaths property to the config which causes the page to be skipped in future rollouts - not sure this is the same for cq:BlueprintSyncConfig as well.
I have a postgresql database which I keep updated using Osmosis. Osmosis can write to two different database schemas, named Simple and Snapshot. There are not that much different from the database Geoserver uses, But I can't make Geoserver use it perfectly.
The main problem seems to be the way tags are stored in those DBs. I can add the nodes layer and display it with that default Points style, but as soon as I use a "ogc:Filter" in my style to filter the nodes by their "place" tag, the WMS is broken and does not respond (says: The requested Style can not be used with this layer. The style specifies an attribute of place and the layer is: TestDB:nodes)
Is there anyway to make GeoServer understand that one of those shemas, or make Osmosis update to the DB GeoServer knows?
This is a case for using TRIGGERs to manage the integration. The two programs use two different schemas. You can CREATE TRIGGERs in the database which ensure that data written to one application is made available to another. Another option is you can set one or both to use VIEWs populated in part by the other application. In PostgreSQL, a VIEW can have triggers attached so these are not really
This is, in any case, a potentially large project so rather than offering sample code, I will offer a general outline of what sorts of things you need to think about.
Are these generally applicable? If so do you want to start an open source integration project?
Are both of these read-only workloads? Does data ever update? In general, if you are going to use views, updates pose the most concerns, so you want to run the views on the side not doing the updates if such is the case.
What is the write model of both sides? Insert/Update? Append only? Static data? What data do you have to "replicate" between the schemas?
Once you have those answers it should be relatively straightforward to get started and ask for help (either as an open source project or here) where you get stuck.
I am trying to follow the Wizard pattern in Forms sample provided with Play (https://github.com/playframework/Play20/blob/master/samples/java/forms/app/views/wizard/form1.scala.html).
This approach looks okay when the number of fields are static. But, how do I deal with this when the fields are dynamic. e.g. if there are going to be multiple profiles that a user can create in step2, how do I represent it on this page?
Also, does it mean that every page of my wizard will have to know about all the controls on rest of the pages, and make those hidden? There must be a better approach to solve this problem. Can you pls help?
I had similar problem while working with wizards. I solved it by decoupling my DB models from UI models. e.g. at DB level, I have one model that represents a whole car. At UI layer, I created multiple models that represent parts of the car e.g. wheels, seats, doors etc.
In the UI wizard, I use the UI models. So, at any given step my wizard step only needs to know about the parts that it operates on. I can apply the validation constraints such as #required etc. on these models. If form validation on part succeeds, I would update the DB model with that information. HTH.