Skip to main content

Home/ BI-TAGS/ Group items tagged tables

Rss Feed Group items tagged

cezarovidiu

How to join GL tables with XLA (SubLedger Accounting) tables - 0 views

  • GL_JE_BATCHES (je_batch_id)                                   => GL_JE_HEADERS (je_batch_id) GL_JE_HEADERS (je_header_id)                                  => GL_JE_LINES (je_header_id) GL_JE_LINES (je_header_id,  je_line_num)                      => GL_IMPORT_REFERENCES (je_header_id, je_line_num) GL_IMPORT_REFERENCES (gl_sl_link_table, gl_sl_link_id)        => XLA_AE_LINES (gl_sl_link_table, gl_sl_link_id) XLA_AE_LINES (application_id, ae_header_id)                   => XLA_AE_HEADERS (application_id, ae_header_id)  XLA_AE_HEADERS (application_id, event_id)                     => XLA_EVENTS (application_id, event_id)    XLA_EVENTS (application_id, entity_id)                        => XLA.XLA_TRANSACTION_ENTITIES (application_id, entity_id) 
  • The source_id_int_1 column of xla.xla_transaction_entities stores the primary_id value for the transactions. You can join the xla.xla_transaction_entities table with the corresponding transactions table for obtaining additional information of the transaction. For e.g you join the xla.xla_transaction_entities table with ra_customer_trx_all for obtaining receivables transactions information or with mtl_material_transactions table for obtaining material transactions information. The entity_id mappings can be obtained from the XLA_ENTITY_ID_MAPPINGS table
  •  
    "How to join GL tables with XLA (SubLedger Accounting) tables"
cezarovidiu

16.4.2. Replication Compatibility Between MySQL Versions - 0 views

  • MySQL supports replication from one major version to the next higher major version. For example, you can replicate from a master running MySQL 4.1 to a slave running MySQL 5.0, from a master running MySQL 5.0 to a slave running MySQL 5.1, and so on.
  • However, one may encounter difficulties when replicating from an older master to a newer slave if the master uses statements or relies on behavior no longer supported in the version of MySQL used on the slave. For example, in MySQL 5.5, CREATE TABLE ... SELECT statements are permitted to change tables other than the one being created, but are no longer allowed to do so in MySQL 5.6 (see Section 16.4.1.4, “Replication of CREATE TABLE ... SELECT Statements”).
  • Important It is strongly recommended to use the most recent release available within a given MySQL major version because replication (and other) capabilities are continually being improved. It is also recommended to upgrade masters and slaves that use early releases of a major version of MySQL to GA (production) releases when the latter become available for that major version.
cezarovidiu

Enkitec : Products : Plug-ins : Plug-in Details - 0 views

  • Plug-in Details: CLOB Load Have you ever used a Rich Text Editor item type in APEX? Has the amount of text ever exceeded 32k? Choke! And so begins the search for a work around... While work arounds exists they can be a little difficult find and repeat each time you need them. We designed the Enkitec CLOB Load plug-in to simplify and standardize how large amounts of text are moved from the database to APEX page items and then back to the database.
cezarovidiu

Convert VirtualBox (vdi) hard drive image to VMWare (vmdk) format » MikeBeach... - 0 views

  • Example (Windows):

    1"c:Program FilesOracleVirtualBoxVBoxManage.exe" clonehd "Win XP.vdi" xp.vmdk  --format vmdk --variant standard
  • Example (Linux):

    1VBoxManage clonehd "Win XP.vdi" xp.vmdk  --format vmdk --variant standard
  • Next, open VMWare and select Create a new virtual machine Select “I will install the operating system later” Make your OS selection about the OS that’s currently on the vmdk you will be using. (The guest OS, not the host OS). Later on, you will have the option to use an existing vmdk image as your virtual hard drive. Do so. You should now be able to finish setup and boot your converted disk image.
  • ...1 more annotation...
  • Absolute path to VBoxManage is necessary unless it’s in the Windows $PATH.
cezarovidiu

OBIEE Consultant Jobs, Average Salary for OBIEE Consultant Jobs - 0 views

  •  
    "The first part of the table below looks at the demand for the OBIEE Consultant role in IT jobs advertised across the UK. Included is a guide to the average salaries offered in IT jobs that have cited OBIEE Consultant in their job title over the 3 months to 7 February 2013 with a comparison to the same period in the previous 2 years. The second part of the table is for comparison and provides aggregates for all Job Titles currently tracked."
cezarovidiu

- Process Type Plugin - SaveToDisk - 0 views

  • Save file to file system rather than into table
  •  
    Save file to file system rather than into table
cezarovidiu

Moving Sugar to Another Server - SugarCRM Support Site - 0 views

    • cezarovidiu
       
      japtone   Senior Member Join Date Nov 2010 Posts 49  Re: Transferring SugarCRM to a new server If you're using Linux try to have the same version of PHP, Apache, and DB (MySQL for instance) in order to avoid compatibility issues. In your production server tar up the sugarcrm root directory, transfer it to the new server and untar wherever your new root directory will be.  Next take a db dump of your database, transfer it to the new server and do a restore. Make sure apache is configured on the new server to point to the root of sugarcrm and start it up.  Make sure to modify config.php to account for any change in paths and hostname.  that's what I've found to be the easiest way to 'clone' sugar.
  • mysqldump -h localhost -u [MySQL user, e.g. root] -p[database password] -c --add-drop-table --add-locks --all --quick --lock-tables [name of the database] > sqldump.sql
  • Extract the Database
  • ...5 more annotations...
  • Copy Filesystem Copy all your files to the new server.  This can be done simply by locating the root directory on your old instance and copy and pasting it to the new server location.
  • Import Database Import the mysql database into the new server.  Here's how you would restore your custback.sql file to the Customers database. mysql -u sadmin -p pass21 Customers < custback.sql Here's the general format you would follow: mysql -u [username] -p [password] [database_to_restore] < [backupfile]
  • Check Files and Permissions Check Config.php Open <sugarroot/config.php> and make sure that all settings still apply to the new server, such as: array ( 'db_host_name' => 'localhost', 'db_user_name' => 'root', 'db_password' => 'PASSWORD', 'db_name' => 'DATABASE_NAME', 'db_type' => 'mysql', ), 'site_url' =>, etc...
  • Check htaccess Open <sugarroot/.htaccess> and ensure that the new server URLs are used correctly.
  • Check Permissions Check that the permissions are correct on the new server. That is the entire custom and cache directories (and all the sub directories) in addition to the config.php file are owned and writable by the user that runs the application on the server.
cezarovidiu

[Tutorial] VLOOKUP questions and answers (View topic) * OpenOffice.org Community Forum - 0 views

  • Summary: Check Search whole cells and uncheck Regular expressions
  •  
    "Very important: Two of the options (OpenOffice.org > Preferences on a Mac, Tools > Options on other platforms) affect several functions, including VLOOKUP. Both of these options are in the Calc > Calculate section: Search criteria = and <> must apply to whole cells - If you uncheck this option text searches in VLOOKUP can match a substring of the values in the table so in the example a search for B will find B+. You almost certainly want to enable this option so that an exact match must occur. Enabling the option also makes your VLOOKUP formulas compatible with Excel. Enable regular expressions in formulas - Unless you understand what "regular expressions" are (see Help) and unless you specifically want to use them in your spreadsheet, you will want to uncheck Enable regular expressions in formulas because this option can make VLOOKUP difficult to use. Unchecking the option also makes your VLOOKUP formulas compatible with Excel. The questions below address what happens if you enable this option. Summary: Check Search whole cells and uncheck Regular expressions"
cezarovidiu

Entity Attribute Value in Magento - Magento tutorial lesson 19 - Magento - 0 views

  • Entity Attribute Value
  • number of attributes (properties and parameters) which can be used to describe them are potentially vast, but the number of&nbsp;attributes&nbsp;which will actually apply to a given&nbsp;entity are relatively modest.
  • sparse matrix
  • ...11 more annotations...
  • vertical database
  • open schema
  • more complex queries
  • Entity: The entity represents Magento data items
  • Attribute: The attributes represent data items that belong to an entity.
  • Value: The value is the simplest to understand as it is simply a value linked to an attribute
  • 1.1. What is Entity Attribute Value
  • 1.2. Entity table structure
  • At the very least, the attribute definitions table would contain the following columns: an attribute ID, attribute name, description, data type, and columns assisting input validation, e.g., maximum string length and regular expression, set of permissible values, etc.
  • The attribute or parameter:
  • The entity
cezarovidiu

OBIEE 11G configuration's Maximum Limits | OBIEE by Shiva Molabanti - 0 views

  • OBIEE 11G configuration’s Maximum&nbsp;Limits
  • Be careful when you are setting the Cache attribute values. For implementations under significant stress increase this value to 1000 or a higher value as mentioned below. The main factor that affects its size is the memory consumption. Maximum range of Global Cache storage size (in Megabytes) to use when BI Server cache is enabled – 9,999 Maximum number of cache entries; used to tune cache performance – 99,999 Maximum size (in Megabytes) for individual server cache entries – 9,999 Data Display/Download: The Maximum no.of rows that we can set to display/download in the Table/Pivot views. But keep in mind OBIEE is NOT meant to be a tool to extract huge amounts of information. Maximum number of rows of data to include in Table views – 9,999,999 Maximum number of rows of data to include in delivered emails – 9,999,999 Maximum number of rows to export to excel – 9,999,999
  • In OBIEE 11G, The instanceconfig.xml file stores the OBI PS configuration settings. Many configuration settings are available in EM and those should be controlled within EM only unless if we disabled centralConfigurationEnabled option. If any particular setting is not available in EM then we should make changes directly in instanceconfig.xml file. There are many situations where we need to change the default OBI configuration settings to a new value.
cezarovidiu

Understanding the Oracle REST Data Services Configuration Folder | cdivilly - 0 views

  • java -jar ords.war configdir /tmp/wls/conf
  • Configure ORDS by doing: 1java -jar apex.war
  •  
    "java -jar apex.war"
cezarovidiu

Rittman Mead Consulting - The Changing World of Business Intelligence - 0 views

  • Schema on write This is the traditional approach for Business Intelligence. A model, often dimensional, is built as part of the design process. This model is an abstraction of the complexity of the underlying systems, put in business terms. The purpose of the model is to allow the business users to interrogate the data in a way they understand.
  • The model is instantiated through physical database tables and the date is loaded through an ETL (extract, transform and load) process that takes data from one or more source systems and transforms it to fit the model, then loads it into the model.
  • The key thing is that the model is determined before the data is finally written and the users are very much guided or driven by the model in how they query the data and what results they can get from the system. The designer must anticipate the queries and requests in advance of the user asking the questions.
  • ...3 more annotations...
  • Schema on read Schema on read works on a different principle and is more common in the Big Data world. The data is not transformed in any way when it is stored, the data store acts as a big bucket. The modelling of the data only occurs when the data is read. Map/Reduce is the clearest example, the mapping is the understanding of the data structure. Hadoop is a large distributed file system, which is very good at storing large volumes of data, this is potential. It is only the mapping of this data that provides value, this is done when the data is read, not written.
  • New World Order So whereas Business Intelligence used to always be driven by the model, the ETL process to populate the model and the reporting tool to query the model, there is now an approach where the data is collected its raw form, and advanced statistical or analytical tools are used to interrogate the data. An example of one such tool is R.
  • The driver for which approach to use is often driven by what the user wants to find out. If the question is clearly formed and the sources of data that are required to answer it well understood, for example how many units of a product have we sold, then the traditional schema on write approach is best.
cezarovidiu

Filling a Critical Role in Business Today: The Data Translator - Microsoft Business Int... - 0 views

  • a lot of articles calling data scientists and statisticians the jobs of the future
  • there are more immediate needs that, when addressed, will have a much greater business impact.
  • Right now we have huge opportunities to make the data more accessible, more “joinable” and more consumable. Leaders don’t want more data – they want more information they can use to run their businesses.
  • ...5 more annotations...
  • Every company has hundreds of millions of records about their sales, expenses, employees and so on, with dozens of insights yet to be discovered through simple comparison or triangulation of relevant data.
  • Why don’t we focus on this? I think because it’s very difficult to do – being successful in this “data translator” role requires a unique set of skills and knowledge, the combination of which I call the BASE skillset: Business understanding Ability to synthetize and simplify Storytelling skills Expertise in data visualization
  • Business Understanding This one seems obvious, but it doesn’t mean simply understanding the financials of a business. Rather, it means truly knowing the operational details, the incentives, the install base, market growth, penetration, the competition, etc. An analyst can’t just know the technical aspect of a report or the math behind the numbers, but what is truly driving a pattern in terms of product quality, competition, incentives and/or offerings. The best analysts are able to mathematically isolate the key levers of a trend and then suggest actions to react to or take advantage of those trends. Ability to Synthetize and Simplify This is, in my opinion, the most underrated and underappreciated skill. Combing through thousands of data points and netting out 3-4 key issues in under 10 minutes, and then communicating these to a group of execs with very different analytical skills, is truly difficult. The key is to make it simple but not simplistic, which means you still capture the complexity even as you get to the few core insights. It requires a very thorough effort to gather all the relevant information before categorizing, prioritizing and deciding if it is significant. After a while, you become an expert and can sniff things out quickly. At the same time, there is the danger of missing anomalies when you jump to conclusions based only on a summary look.
  • Storytelling Skills There are stages that should be followed when explaining complex ideas, something data translators are frequently expected to do. The best storytellers start by giving context and trying to couple the current discussion to something the audience already knows, ensuring the story is well structured and connected. We have to move from a “buffet style” business review with thousands of numbers packed in tables to a layered approach that will guide the audience to focus first on the most relevant messages, diving deeper only when necessary. Minto Pyramid Principles, which are built around a process for organizing thought and communication, are helpful in making sure you really focus on what is important and relevant, versus being obsessed in telling every fact. Expertise in Data Visualization I am glad to finally see so much focus on Information Visualization and I believe this is correlated to the explosion of data. Traditional methods of organizing data do not facilitate an intuitive understanding of key information points or trends. For instance, the two examples below contain data on car sales across the U.S. The first, an alphabetized list, is much less intuitive than the second, which shows those sales on a map in Power View. With Power View, right away you can identify the states with the highest sales: CA, FL, TX, NY. (Workbook available here)
  • There is no better way to see patterns or trends than data visualization, making expertise in this area – both technical and analytical – critical for data translators.
cezarovidiu

You Probably Need Parallel Except When You Don't - 0 views

  • f you are running a large Oracle data warehouse you should be using parallel
  • Like all tools you have to use parallel correctly; no more would we think of using a wrench to hammer a nail then should you think parallel is the answer to all performance problems. Sometimes parallel will make things worse, sometimes parallel will make performance less predictable.
  • Parallel introduces additional work to a query, simplistically we need to: split the query into multiple parallel processes, execute them, wait for the processes to complete and finally coordinate the results. This all takes time to do. Our time saving comes from being able to process multiple smaller chunks of data simultaneously. If the time to execute the step in parallel is not significantly faster than doing it without parallel then the additional overhead may make parallel processing a slower option; this is typically the case with small tables where a full tablescan or an indexed access is fast. Use too few parallel processes and we will not gain much in performance, too many and we risk starving the database of resource for other work or even slow our own process as it waits for resource. If you have implemented some form of CPU resource management on your system you may find that you experience delays as your parallel slaves ‘wait their turn’
cezarovidiu

mysql - Best of MyISAM and InnoDB - Database Administrators - 0 views

  • Some people can make the table's row format FIXED using ALTER TABLE mydb.mytb ROW_FORMAT=Fixed; and can get a 20% increase in read performance without any other changes. This works and works effectively FOR MyISAM. This will not produce faster results for InnoDB because ... that's right ... you must consult the gen_clust_index each time.
cezarovidiu

Magic Quadrant for Business Intelligence and Analytics Platforms - 0 views

  • Integration BI infrastructure: All tools in the platform use the same security, metadata, administration, portal integration, object model and query engine, and should share the same look and feel. Metadata management: Tools should leverage the same metadata, and the tools should provide a robust way to search, capture, store, reuse and publish metadata objects, such as dimensions, hierarchies, measures, performance metrics and report layout objects. Development tools: The platform should provide a set of programmatic and visual tools, coupled with a software developer's kit for creating analytic applications, integrating them into a business process, and/or embedding them in another application. Collaboration: Enables users to share and discuss information and analytic content, and/or to manage hierarchies and metrics via discussion threads, chat and annotations.
  • Information Delivery Reporting: Provides the ability to create formatted and interactive reports, with or without parameters, with highly scalable distribution and scheduling capabilities. Dashboards: Includes the ability to publish Web-based or mobile reports with intuitive interactive displays that indicate the state of a performance metric compared with a goal or target value. Increasingly, dashboards are used to disseminate real-time data from operational applications, or in conjunction with a complex-event processing engine. Ad hoc query: Enables users to ask their own questions of the data, without relying on IT to create a report. In particular, the tools must have a robust semantic layer to enable users to navigate available data sources. Microsoft Office integration: Sometimes, Microsoft Office (particularly Excel) acts as the reporting or analytics client. In these cases, it is vital that the tool provides integration with Microsoft Office, including support for document and presentation formats, formulas, data "refreshes" and pivot tables. Advanced integration includes cell locking and write-back. Search-based BI: Applies a search index to structured and unstructured data sources and maps them into a classification structure of dimensions and measures that users can easily navigate and explore using a search interface. Mobile BI: Enables organizations to deliver analytic content to mobile devices in a publishing and/or interactive mode, and takes advantage of the mobile client's location awareness.
  • Analysis Online analytical processing (OLAP): Enables users to analyze data with fast query and calculation performance, enabling a style of analysis known as "slicing and dicing." Users are able to navigate multidimensional drill paths. They also have the ability to write back values to a proprietary database for planning and "what if" modeling purposes. This capability could span a variety of data architectures (such as relational or multidimensional) and storage architectures (such as disk-based or in-memory). Interactive visualization: Gives users the ability to display numerous aspects of the data more efficiently by using interactive pictures and charts, instead of rows and columns. Predictive modeling and data mining: Enables organizations to classify categorical variables, and to estimate continuous variables using mathematical algorithms. Scorecards: These take the metrics displayed in a dashboard a step further by applying them to a strategy map that aligns key performance indicators (KPIs) with a strategic objective. Prescriptive modeling, simulation and optimization: Supports decision making by enabling organizations to select the correct value of a variable based on a set of constraints for deterministic processes, and by modeling outcomes for stochastic processes.
  • ...7 more annotations...
  • These capabilities enable organizations to build precise systems of classification and measurement to support decision making and improve performance. BI and analytic platforms enable companies to measure and improve the metrics that matter most to their businesses, such as sales, profits, costs, quality defects, safety incidents, customer satisfaction, on-time delivery and so on. BI and analytic platforms also enable organizations to classify the dimensions of their businesses — such as their customers, products and employees — with more granular precision. With these capabilities, marketers can better understand which customers are most likely to churn. HR managers can better understand which attributes to look for when recruiting top performers. Supply chain managers can better understand which inventory allocation levels will keep costs low without increasing out-of-stock incidents.
  • descriptive, diagnostic, predictive and prescriptive analytics
  • "descriptive"
  • diagnostic
  • data discovery vendors — such as QlikTech, Salient Management Company, Tableau Software and Tibco Spotfire — received more positive feedback than vendors offering OLAP cube and semantic-layer-based architectures.
  • Microsoft Excel users are often disaffected business BI users who are unable to conduct the analysis they want using enterprise, IT-centric tools. Since these users are the typical target users of data discovery tool vendors, Microsoft's aggressive plans to enhance Excel will likely pose an additional competitive threat beyond the mainstreaming and integration of data discovery features as part of the other leading, IT-centric enterprise platforms.
  • Building on the in-memory capabilities of PowerPivot in SQL Server 2012, Microsoft introduced a fully in-memory version of Microsoft Analysis Services cubes, based on the same data structure as PowerPivot, to address the needs of organizations that are turning to newer in-memory OLAP architectures over traditional, multidimensional OLAP architectures to support dynamic and interactive analysis of large datasets. Above-average performance ratings suggest that customers are happy with the in-memory improvements in SQL Server 2012 compared with SQL Server 2008 R2, which ranks below the survey average.
  •  
    "Gartner defines the business intelligence (BI) and analytics platform market as a software platform that delivers 15 capabilities across three categories: integration, information delivery and analysis."
cezarovidiu

Angel's BI Blog: Excel Data Explorer and the Twitter Search API - 0 views

  • its a very powerful query engine/data transformation tool.
  • what's fascinating is its ability to load web data from static html pages, tables on web pages, web services, etc. I had a need to search recent activity on Twitter and decided to test drive Data Explorer in Excel.
1 - 20 of 21 Next ›
Showing 20 items per page