The DFS browser cannot browse anything else but a distributed file system - eclipse

I have installed hadoop 1.2.1 (all on single machine). I am getting an error in eclipse that
'The DFS browser cannot browse anything else but a distributed file system'.
Also tried with changing core-site.xml value
<value>hdfs://localhost:50040</value>
to
<value>hdfs://[your IP]:50040</value>

Please make sure your NameNode is running at port 50040. If you have not changed it then it would be running on port 9000. Also, make sure that your Hadoop plugin contains all the required jars in it.

Related

Typo3 CMS throws Error after successful installation

Once I finish installing Typo3 locally on my system using latest xampp on PHP 8 version. Website stops working. Installation gets completed successfully, but once installation is completed, whole website stops working. This is strange. Every time I tried installing latest Typo3 version, 11.5, same issue happens.
Depending on the kind of installation you're likely missing the correct server configuration. Usually it's done by a .htaccess file in the web-root of the installation.
An example .htaccess file, which in many installations never even needs adjustments, is located inside your installation, you just have to copy it in the webroot and rename it to .htaccess. It's located here:
typo3\sysext\install\Resources\Private\FolderStructureTemplateFiles\root-htaccess
In your Screenshot, you are using Port 81 (localhost:81).
I suggest using the default Port 80 and "127.0.0.1" instead of "localhost" (http://127.0.0.1).
Using Port 81 would need changes in your firewall configuration in many cases.
Additionally, your folder structure seems wrong ("/typo3_src-11.5.12/typo3/login?...").
Your backend should be reachable under "/typo3/" or "/typo3/index.php".
If you are using a legacy installation (without composer), the folder /typo3_src-11.5.12/ should be one level above your web server's document root, as stated here:
https://docs.typo3.org/m/typo3/tutorial-getting-started/main/en-us/Installation/LegacyInstallation.html

GridGain. Local node's library list differs from remote node's, leveldbjni-all-1.1.jar

When I try to use gridGain on one machine, I run a ggstart.bat and run app on eclipse, I added all .jar files from gridGain sourŅe folder, but I see on console this when start:
Local node's library list differs from remote node's
<leveldbjni-all-1.1.jar> vs. <not found>
(in ecsiple's console conversely - not found vs level...)
Only this file from all, I try delete this, in doesn't work, try to rename, message change to <newFile> vs. <not found>
I try to found about this file, found only "download leveldbjni-all-1.1.jar" and something different size of file, but no about this file.
May be someone knows how to fix it or something about this file?
Version of gridGain is 5.3.0
The jar your are talking about is a LevelDB library, used in one of GridGain swapspace implementations. Please check if you are running the same version of GridGain remotely and from Eclipse, and that the classpath is equal.

DFS locations in Eclipse Europa after accessing Hadoop in a VM

I am very new to hadoop. I need to install it and play around with samples.
SO i referred this tutorial . I have installed Sandbox given in that tutorial. I need to configure ECLIPSE in windows mentioning VM location as specified in the image below, which is given in the tutorial.
I have installed eclipse europa and hadoop plugin.
Then in Map/Reduce Locations i gave VM Ip for host name, Linux user name in UserName and 9001 in Map/Reduce port and 9000 in DFS port.
In Advanced Tab I have set value to the mapred.system.dir as /hadoop/mapred/system
and there is no hadoop.job.ugi to give username.
After i click ok, I couldn't get HDFS file system under my DFS locations in ECLIPSE.
Please help me on this
I also got the same problem. The problem here is not related to hadoop configuration but eclipse. To fix this, go to "\workspace\.metadata\.plugins\org.apache.hadoop.eclipse\locations". Here open the XML file and just add the property "hadoop.job.ugi" with value "hadoop-user,ABC" and then restart your eclipse. It worked for me.
I tried by giving just one value i.e. without ABC but it dint work and I dont know the significance of this comma separated value but since I have just started the tutorial I hope to get this answer soon :)
I too ran into the same issue. I installed RedHad cgywin (openssh and openssl packages) and updated the "Path" environment variable with a path to cgywin/bin (c:\rhcygwin\bin). Then my Eclipse DFS location was able to connect to Hadoop on the Virtual Machine. Once that was successful I saw the option "hadoop.job.ugi".
http://v-lad.org/Tutorials/Hadoop/00%20-%20Intro.html describes installing cgywin.
Note: I am running the Hadoop VM on Windows Vista.
I spawned Eclipse from within Cygwin and it worked fine for me (i.e. I could see the "hadoop.job.ugi" parameter). Also, I didn't make any changes to my PATH environment variable.

Hadoop location in Eclipse (Error: null)

I've downloaded hadoop from Yahoo tutorial, started linux VM with hadoop and in eclipse created new DFS location (entered IP of my VM, Map/Reduce master port 9001, DFS port 9000.
But in the node I got the error "Error:null."
What I'm doing wrong?
I'm using Eclipse Europe 3.3.1 and Hadoop 0.18.0.
Thanks for helping.
You should check the additional properties that you are able to configure. I was also getting this problem but the actual problem was that hadoop.job.ugi property was not available to be setup.To fix this, go to "\workspace\.metadata\.plugins\org.apache.hadoop.eclipse\locations". Here open the XML file and just add the property "hadoop.job.ugi" with value "hadoop-user,ABC" and then restart your eclipse. It worked for me.
I've configured properly eclipse when had installed it in my VM where hadoop was. But due to incompatible new versions of eclipse and eclipse hadoop plugin I refused to use it, because seems to me that using this plugin doesn't have any good benefits.
I was facing the same problem, i thinks the problem because there is no folder or files in the HDFS,i solved with these steps:
hadoop dfs -mkdir /name of folder
hadoop fs -ls /

Eclipse: Glassfish and Tomcat don't work

I have reinstalled java JDK 6.0.25,(after some problems installing java, finally succeeded). I have set paths (JAVA_HOME points to C:\Program Files\Java\jdk1.6.0_25 and in Path also added C:\Program Files\Java\jdk1.6.0_25\bin).
In eclipse I have pointed everything to this new JDK, also in servers. When I have installed Glassfish 3.1, after installation the domain wasn't created. So when I add this server in Eclipse, the domain wasn't found.
So I did extract an ZIP version of Glassfish and in Eclipse configure the server to this ZIPPED Glassfish. Also I have installed tomcat 7.0.14 again and add to eclipse but both did not work.
Main problem: When I start Tomcat or Glassfish in Eclipse they don't start. Tomcat is starting but after 45 seconds it says, that the time is run out. I have read the forums to increase starting time, I have putted it to 240 sec, but it does not work, because it pop out error in which sazs, that the port is already used by a different server.
By Glassfish, when I start it, its starting, but after few minutes gives me error: Failed to start. It looks like some ports or something is blocking this servers. I have turned off firewall also and antivirus but nothing.
I have looked all the forum about "not starting Glassfish or Tomcat" but didn't found the solution. I don't want to reinstall windows for this.
So if anyone know anything about this problem, please HELP.
Update:
I've tried to change the ports in domain.xml and checked with netstat -b (everything was established), but the server (Glassfish) didn't start. It shows an error "Server GlassFish Server Open Source Edition 3 (Java EE 6) at localhost failed to start.". I've also tried install "built in Eclipse" Glassfish from marketspace and from "Install new software" from eclipse, but it comes out with the same error. Currently, for needs I'm using XP Mode in Windows 7, there servers work, without any problems.
It seems your installation odyssey messed up your system.
But you should get it working without reinstalling Windows. Maybe the fragment of an old server installation blocks the ports Glassfish and Tomcat need for operation.
Try the following:
Check which ports your "new" servers try to access (should be something like 8080 and should be written in server logs during startup)
Use netstat -b (from cmd started with admin rights) to check which application blocks this / these ports (could be java.exe if there is an old installation still hanging)
Remove the old installation (uninstall or disable service)
Try again to start your new server(s).
Try tell to GF or TC to bind the default web listener other port than 8080, because this is a root of problem.
I doesn't know well TC, but in GF case, this is very easy. Simply edit $GLASSFISH_HOME/domains/domain1/config/domain.xml and find something similar:
<network-listener port="8080" protocol="http-listener-1" transport="tcp" name="http-listener-1" thread-pool="http-thread-pool"></network-listener>
And replace 8080 with other port (but not same with TC). In your case I could choose 8888 which is easy to remember but you can pretty sure this is differs from TC's port.
Note: Edit this file before you start GF!