I am a beginner on the Scorm. I want to access scorm content from a different server. a video/image is stored on s1 instance and accessing the content from an LMS via scorm which is hosted on a different instance, s2.
What needs to be modified in the scorm to make it accessible to the cloud-hosted scorm content?
I changed the content location path from imsmanifest.xml file. But still not access the cloud content.
The following steps are followed to update a Scorm content with cloud content.
Took a scorm package.
Unzip the scorm package and changed an mp4 video content with cloud-hosted URL. Screenshot:
Zip the scorm and upload on a Moodle LMS.
Experiencing this error:
It seems it is still taking the old path.
Listing the file file in SCORM manifest file is just to tell the LMS what files to expect in the package (and I'm not even sure if Moodle's SCORM implementation uses that information at all, or just accepts what's in the package).
For the video hosted outside the package to be accessed by the user, you need to use the full path in the HTML or javascript code used to display the video.
This isn't actually a SCORM issue, it's a run-of-the-mill HTML web page issue.
Related
I'm developing a web app for in-house use and I'm looking for a better way to display PDFs.
I've played around with Adobe's 'Work with Local File' example from GitHub, Adobe GitHub Example, and it works great using the file picker to display a PDF. Is it possible with Adobe's PDF Embed API to take a file located on a local file share and display the PDF?
I'm thinking I need to create a file promise but I'm not sure how to create that.
Unless you can make a network request to load the PDF, the answer is no. Browsers generally can't read from local files unless a user action actually picks the file. If your local share can be made accessible via HTTP, then you would be good to go.
I tried different method to include the content of an external webpage by url in dita files without success. Is it possible to do it? I can add references to my dita files:
<related-links>
<link href="http://example.com" format="html" scope="external"/>
</related-links>
however I am not able to ddisplay the content of the page in oXygen.
tl;dr: It depends on the external sources and on your publication targets. If you want to publish to PDF, your content needs to be available offline.
What do you exactly mean with content? You can use external sources, e.g. images, if you just publish to html5. Just use the image URL in the href. If you publish to pdf, it would not work. The DITA-OT does not download things for you. Most websites are dynamic, so content is not very static so it may not work as you might expect. Maybe the external website has an API to gather the data you need. Or maybe it would be safer to use the depend.preprocess.pre extension point to hook in an Ant Target to download the external sources via <get> or curl. Another approach would be dita-glass for on-the-fly conversion. But, it depends...
We're considering using Adobe Experience Manager for an upcoming project. But we have a number of pre-made static pages we'd like to import into this project. What would be the best way to 'import' these pages into an AEM project?
Note: these pages are HTML that may contain some CSS and JS.
You can serve static HTML, CSS and JS files from AEM. There are multiple ways of getting your files in AEM including but not limited to:
1) Through CRXDE, goto Create > Create File... name your HTML file and save. On the jcr:content subnode, double click on the jcr:data binary property, upload your HTML file and click save.
2) cURL your files into AEM. See the documentation on the SlingPostServlet. You can write a script in Bash, Python or your language of choice to loop over your files and POST to the AEM instance.
3) Go to CRXDE Package Manager, create a simple package, download it and unzip the file. Examine the contents of the zip including the .content.xml files and the /META-INF/vault/filter.xml file. Add your HTML files and update the package filters, zip up the files, upload the package to AEM through the CRXDE Package Manager and install.
Adding to what "nateyolles" has already mentioned,
Adobe Experience Manager provides several options (direct/Indirect) for either importing or exporting data (within the AEM tool they are labeled importers and are found under the tools area). Lets look at each option so you can decide which one will best service your current needs.
1) Site Importer – Allows you to import either a web page or an entire website into AEM. For site imports, you will need to need to specify the project title, site URL (of page or site to import), thumbnail image for the template, the sitemap (for the pages in CQ using the template) and whether or not you want to overwrite (if page or site exist in AEM)
"Use of Site Importer tool"
Link:- https://docs.adobe.com/docs/en/aem/6-2/develop/dev-tools/site-importer.html
2) Feed Importer – The feed import option within Adobe Experience Manager provides the ability to import content via a feed (whether that be an RSS feed or an Atom feed). This can be extremely beneficial to blog creators, like myself, who might want to bring in the content of their blog (from say WordPress, Drupal or another CMS) into an AEM environment.
3) Offline Importer – This functionality, within Adobe Experience Manager, provides the ability to import content from Microsoft Office Word document files into AEM templates (ending with the file extension of .doc). Please note that you should reach out to Adobe’s consulting team for asssistance on this so that they can set up the inputs in the most optimal way to receive your offline data.
4) Import from older version of CQ – This is a pretty straightforward way to import or upgrade from an older version of CQ. If you are familar with AEM, you can stop existing instance, if you have a stand alone instance of AEM replace the jar file, and restart
If you have an application server instance, you will follow these steps:
Stop existing instance, replace the war file and restart.
5) Bulk Editor – This feature provides a multitude of functionality within AEM. For starters, you can make large changes to content at once. Secondly, you can export content to CSV (Microsoft Excel) files.
Kautuk Sahni
(Adobe AEM Community)
I developed a simple app in Unity for a web site. How can I add my app to my wix site?
I tried finding on the web articles on how to do so but found nothing useful.
A web build will provide you with an HTML and a .unity3d file. If your host allows you to embed HTML, take the relevant sections out of the HTML file and upload your .unity3d file.
If your host does not allow you to upload the required .unity3d file, you'll have to host it yourself. This can be on your own server, on a Dropbox, or wherever you find appropriate. This does require a small change in the generated JavaScript.
Inside your HTML file you'll find a line like
u.initPlugin(jQuery("#unityPlayer")[0], "myAwesomeGame.unity3d");
Of course the exact name depends on however your build was named. If you don't host it in the same location as the HTML, you'll need to update this.
Say you're uploading it to your public Dropbox folder. In that case you can get a public link for the Unity3D file. Replace the call above with something like
u.initPlugin(jQuery("#unityPlayer")[0], "https://dl.dropboxusercontent.com/u/XXXXXXXX/myAwesomeGame.unity3d");
and you should be good to go. Of course the same works for other hosts.
I have installed TYPO3 on my web server and it use the "Introduction Package" template.
The problem is that I have to work on another TYPO3 online website that use a custom template.
Can I obtain the template of the online website and use it on my local website? (to perform some test in local).
I saw that TYPO3 uses various templating systems and I do not know how this could affect the export\import operation. In Joomla or WP I simply take the template directory and then I copy it into the themes directory but I think that in TYPO3 this operation could be more complex.
The fastest way is to copy these directories:
fileadmin
uploads
typo3conf/ext
and the complete database, then you have "cloned" the original website locally. (assuming that the core stuff is loaded the default way, with symlinks)
Then you just need to edit the database connection stuff and maybe some paths for image magick and so on in the typo3conf/LocalConfiguration.php.