Skip to main content

Home/ BI-TAGS/ Group items tagged Java

Rss Feed Group items tagged

cezarovidiu

Install Oracle Java 7 in Ubuntu 12.10/12.04/11.10/Any Ubuntu or Linux Mint Version ~ No... - 0 views

  • For 64-bit users: wget -O jdk-64bit.tar.gz http://goo.gl/MSzBj wget -O jre-64bit.tar.gz http://goo.gl/yZgjI sudo -s cp -r jre-64bit.tar.gz /usr/local/java sudo -s cp -r jdk-64bit.tar.gz /usr/local/java cd /usr/local/java sudo -s chmod a+x jre-64bit.tar.gz sudo -s chmod a+x jdk-64bit.tar.gz sudo -s tar xvzf jre-64bit.tar.gz sudo -s tar xvzf jdk-64bit.tar.gz
  • sudo nano /etc/profile Add the following lines at the end of file: JAVA_HOME=/usr/local/java/jdk* PATH=$PATH:$HOME/bin:$JAVA_HOME/bin JRE_HOME=/usr/local/java/jre* PATH=$PATH:$HOME/bin:$JRE_HOME/bin export JAVA_HOME export JRE_HOME export PATH
  • Now enter following commands one by one in terminal: sudo update-alternatives --install "/usr/bin/java" "java" "/usr/local/java/jre1.7.0_12/bin/java" 1 sudo update-alternatives --install "/usr/bin/javac" "javac" "/usr/local/java/jdk1.7.0_12/bin/javac" 1 sudo update-alternatives --install "/usr/bin/javaws" "javaws" "/usr/local/java/jre1.7.0_12/bin/javaws" 1 sudo update-alternatives --set java /usr/local/java/jre1.7.0_12/bin/java sudo update-alternatives --set javac /usr/local/java/jdk1.7.0_12/bin/javac sudo update-alternatives --set javaws /usr/local/java/jre1.7.0_12/bin/javaws . /etc/profile
  • ...1 more annotation...
  • Check installed java version java -version
cezarovidiu

Understanding the Oracle REST Data Services Configuration Folder | cdivilly - 0 views

  • java -jar ords.war configdir /tmp/wls/conf
  • Configure ORDS by doing: 1java -jar apex.war
  •  
    "java -jar apex.war"
cezarovidiu

Connecting Infobright and Talend - 1 views

  • These instructions assume that you have Infobright installed and running.   First and foremost, download Talend.  In this example, we will download Talend Open Source Data Integrator v5.0. (http://www.talend.com/download.php)  Once fully installed, download the Talend/Infobright Connector.  Ensure you download the right connector; instructions are on the download page (http://www.infobright.org/Downloads/Contributed-Software/) If you download Talend 4.0+, you’ll want the latest connector For older versions of Talend, you’ll want the 3.7 connector and lower. Once downloaded, perform the following actions: [For Windows] Copy the infobright_jni([_32|_64])bit.dll to C:\Windows\infobright_jni.dll Copy the zipped “tInfobrightOutput” directory to this directory: [Install Root of Talend] \plugins\org.talend.designer.components.localprovider_5.0.1.r74687\components\tInfobrightOutput Copy “infobright-core-3.4.jar” to [Install Root of Talend]\lib\java Running Talend in Windows If using Windows, run talend as Administrator.  If you don’t, you will see odd “Access Denied” or “Accesse Refuse” error messages when trying to use the connector.
  • You need to do some work on these instructions. Version 5 is not like version 4. You must run Talend 5 before the “lib\java\” folder appears.  Once it does appear, it no longer contains the .jar files like version 4; just a file “index.xml” that you have to edit to point to the infobright jar file in the components folder.
cezarovidiu

Installing Hadoop for Fedora & Oracle Linux(Single Node Cluster) | accretion infinity - 0 views

  • Hadoop is a framework written in Java for running applications on large clusters of commodity hardware and incorporates features similar to those of the Google File System (GFS) and of the Map Reduce computing paradigm. Hadoop’s HDFS is a highly fault-tolerant distributed file system and, like Hadoop in general, designed to be deployed on low-cost hardware. It provides high throughput access to application data and is suitable for applications that have large data sets.
  • Some of the Hadoop projects we will talk about are: HDFS : A distributed filesystem that runs on large clusters of commodity machines. Map Reduce: A distributed data processing model and execution environment that runs on large clusters of commodity machines. Pig: A data flow language and execution environment for exploring very large datasets. Pig runs on HDFS and MapReduce clusters. HBase: A distributed, column-oriented database. HBase uses HDFS for its underlying storage, and supports both batch-style computations using MapReduce and point queries (random reads). ZooKeeper: A distributed, highly available coordination service. ZooKeeper provides primitives such as distributed locks that can be used for building distributed applications. Oozie: Oozie is a workflow scheduler system to manage Apache Hadoop jobs.
  • Oracle Linux as the operating system and Hadoop 1.1.2 or 1.2.0
1 - 11 of 11
Showing 20 items per page