Skip to main content

Home/ BI-TAGS/ Group items tagged query

Rss Feed Group items tagged

cezarovidiu

Google Reader (250) - 0 views

  • What this means in practice is that when the BI Server component starts up, it creates and reserves a number of threads in advance, determined by a number of parameters including SERVER_THREAD_RANGE.
  • You can see these threads running and ready to perform tasks for the BI Server component by using a tool such as Process Explorer for Windows
  • Thinking it through a bit, any given single query is, to a certain extent, only really going to use a small part of the total amount of CPUs available on a server, because it’s not the BI Server that runs queries in parallel, it’s the underlying database. For example, a single analysis against a single Oracle Database datasource would only really need a single BI Server thread to handle the query request, but when the underlying database receives the query, it might use a large number of its CPUs to process the query, returning results back to the BI Server to then pass back to the Presentation Server for display to the user.
  • ...2 more annotations...
  • The BI Server wouldn’t have any use for any more query threads, as it can’t really do anything with them – the exception to this being queries that generate multiple physical SQLs, for example to join data from multiple sources together and return a single set of data to the user, for which the BI Server could benefit from a higher CPU count if each of these queries in turn led to lots of threads being used – but two queries, in themselves, don’t neccessarily require two CPUs, because of course the BI Server, and the underlying CPUs, are themselves multi-threaded.
  • To conclude then – all things begin equal, the BI Server should make use of all of the CPUs that the underlying operating system presents to it, with the OS itself deciding what threads are scheduled against which CPUs. In-theory, all CPUs on the server are available to each BI Server component, but each OS is different and it might be worth experimenting if you’re sure that certain CPUs aren’t being used – but this is most probably unlikely and the main reason you’d really consider vertical scale-out of BI Server components is for fault-tolerance, or if you’re using a 32-bit OS and each process can only see a subset of the total overall memory. And, bear in mind that however many CPUs the BI Server has available to it, for queries that send just a single SQL statement down to the underlying database server, adding more CPUs or faster CPUs isn’t going to help as only a single (or so) thread will be needed to send the query from the BI Server to the database, and it’s the database that’s doing all of the work – all that this would help with is compilation and post-aggregation work, and enabling the server to handle a higher number of concurrent users. Invest in a better underlying database instead, sort out your data model, and make sure your data source back-end is as optimised as possible.
cezarovidiu

Download Microsoft Power Query for Excel - Office.com - 0 views

  • Microsoft Power Query is an Excel add-in that enhances the self-service Business Intelligence experience in Excel by simplifying data discovery and access.
  • Power Query enables users to easily discover, combine, and refine data for better analysis in Excel. Power Query includes a public search feature that is currently intended for use in the United States only.
cezarovidiu

Rittman Mead Consulting - The Changing World of Business Intelligence - 0 views

  • Schema on write This is the traditional approach for Business Intelligence. A model, often dimensional, is built as part of the design process. This model is an abstraction of the complexity of the underlying systems, put in business terms. The purpose of the model is to allow the business users to interrogate the data in a way they understand.
  • The model is instantiated through physical database tables and the date is loaded through an ETL (extract, transform and load) process that takes data from one or more source systems and transforms it to fit the model, then loads it into the model.
  • The key thing is that the model is determined before the data is finally written and the users are very much guided or driven by the model in how they query the data and what results they can get from the system. The designer must anticipate the queries and requests in advance of the user asking the questions.
  • ...3 more annotations...
  • Schema on read Schema on read works on a different principle and is more common in the Big Data world. The data is not transformed in any way when it is stored, the data store acts as a big bucket. The modelling of the data only occurs when the data is read. Map/Reduce is the clearest example, the mapping is the understanding of the data structure. Hadoop is a large distributed file system, which is very good at storing large volumes of data, this is potential. It is only the mapping of this data that provides value, this is done when the data is read, not written.
  • New World Order So whereas Business Intelligence used to always be driven by the model, the ETL process to populate the model and the reporting tool to query the model, there is now an approach where the data is collected its raw form, and advanced statistical or analytical tools are used to interrogate the data. An example of one such tool is R.
  • The driver for which approach to use is often driven by what the user wants to find out. If the question is clearly formed and the sources of data that are required to answer it well understood, for example how many units of a product have we sold, then the traditional schema on write approach is best.
cezarovidiu

Rittman Mead Consulting » Blog Archive » Oracle Database Resource Manager and... - 0 views

  • OBIEE, at the BI Server level. lets you define query limits that either warn or stop users from exceeding certain elapsed query times or number of rows returned. Assuming you define a “standard” group for most OBIEE users, you might want to stop them from displaying reports (requests) that return more than 50,000 rows, whilst you might want to warn them if their query takes over five minutes to run.
cezarovidiu

Magic Quadrant for Business Intelligence and Analytics Platforms - 0 views

  • Integration BI infrastructure: All tools in the platform use the same security, metadata, administration, portal integration, object model and query engine, and should share the same look and feel. Metadata management: Tools should leverage the same metadata, and the tools should provide a robust way to search, capture, store, reuse and publish metadata objects, such as dimensions, hierarchies, measures, performance metrics and report layout objects. Development tools: The platform should provide a set of programmatic and visual tools, coupled with a software developer's kit for creating analytic applications, integrating them into a business process, and/or embedding them in another application. Collaboration: Enables users to share and discuss information and analytic content, and/or to manage hierarchies and metrics via discussion threads, chat and annotations.
  • Information Delivery Reporting: Provides the ability to create formatted and interactive reports, with or without parameters, with highly scalable distribution and scheduling capabilities. Dashboards: Includes the ability to publish Web-based or mobile reports with intuitive interactive displays that indicate the state of a performance metric compared with a goal or target value. Increasingly, dashboards are used to disseminate real-time data from operational applications, or in conjunction with a complex-event processing engine. Ad hoc query: Enables users to ask their own questions of the data, without relying on IT to create a report. In particular, the tools must have a robust semantic layer to enable users to navigate available data sources. Microsoft Office integration: Sometimes, Microsoft Office (particularly Excel) acts as the reporting or analytics client. In these cases, it is vital that the tool provides integration with Microsoft Office, including support for document and presentation formats, formulas, data "refreshes" and pivot tables. Advanced integration includes cell locking and write-back. Search-based BI: Applies a search index to structured and unstructured data sources and maps them into a classification structure of dimensions and measures that users can easily navigate and explore using a search interface. Mobile BI: Enables organizations to deliver analytic content to mobile devices in a publishing and/or interactive mode, and takes advantage of the mobile client's location awareness.
  • Analysis Online analytical processing (OLAP): Enables users to analyze data with fast query and calculation performance, enabling a style of analysis known as "slicing and dicing." Users are able to navigate multidimensional drill paths. They also have the ability to write back values to a proprietary database for planning and "what if" modeling purposes. This capability could span a variety of data architectures (such as relational or multidimensional) and storage architectures (such as disk-based or in-memory). Interactive visualization: Gives users the ability to display numerous aspects of the data more efficiently by using interactive pictures and charts, instead of rows and columns. Predictive modeling and data mining: Enables organizations to classify categorical variables, and to estimate continuous variables using mathematical algorithms. Scorecards: These take the metrics displayed in a dashboard a step further by applying them to a strategy map that aligns key performance indicators (KPIs) with a strategic objective. Prescriptive modeling, simulation and optimization: Supports decision making by enabling organizations to select the correct value of a variable based on a set of constraints for deterministic processes, and by modeling outcomes for stochastic processes.
  • ...7 more annotations...
  • These capabilities enable organizations to build precise systems of classification and measurement to support decision making and improve performance. BI and analytic platforms enable companies to measure and improve the metrics that matter most to their businesses, such as sales, profits, costs, quality defects, safety incidents, customer satisfaction, on-time delivery and so on. BI and analytic platforms also enable organizations to classify the dimensions of their businesses — such as their customers, products and employees — with more granular precision. With these capabilities, marketers can better understand which customers are most likely to churn. HR managers can better understand which attributes to look for when recruiting top performers. Supply chain managers can better understand which inventory allocation levels will keep costs low without increasing out-of-stock incidents.
  • descriptive, diagnostic, predictive and prescriptive analytics
  • "descriptive"
  • diagnostic
  • data discovery vendors — such as QlikTech, Salient Management Company, Tableau Software and Tibco Spotfire — received more positive feedback than vendors offering OLAP cube and semantic-layer-based architectures.
  • Microsoft Excel users are often disaffected business BI users who are unable to conduct the analysis they want using enterprise, IT-centric tools. Since these users are the typical target users of data discovery tool vendors, Microsoft's aggressive plans to enhance Excel will likely pose an additional competitive threat beyond the mainstreaming and integration of data discovery features as part of the other leading, IT-centric enterprise platforms.
  • Building on the in-memory capabilities of PowerPivot in SQL Server 2012, Microsoft introduced a fully in-memory version of Microsoft Analysis Services cubes, based on the same data structure as PowerPivot, to address the needs of organizations that are turning to newer in-memory OLAP architectures over traditional, multidimensional OLAP architectures to support dynamic and interactive analysis of large datasets. Above-average performance ratings suggest that customers are happy with the in-memory improvements in SQL Server 2012 compared with SQL Server 2008 R2, which ranks below the survey average.
  •  
    "Gartner defines the business intelligence (BI) and analytics platform market as a software platform that delivers 15 capabilities across three categories: integration, information delivery and analysis."
cezarovidiu

Star Schema Bechmark: InfoBright, InfiniDB and LucidDB - MySQL Performance Blog - 0 views

  • Queries time
  • InfoBright was fully 1 CPU bound during all queries.
  • InfiniDB is otherwise was IO-bound, and processed data fully utilizing sequential reads and reading data with speed 120MB/s. I think it allowed InfiniDB to get the best time in the most queries.
  • ...1 more annotation...
  • LucidDB on this stage is also can utilize only singe thread with results sometime better, sometime worse than InfoBright.
  •  
    "Star Schema Bechmark: InfoBright, InfiniDB and LucidDB"
cezarovidiu

Oracle APEX Interactive report based on PLSQL function | Oracle Application Express - 0 views

  • Unfortunately Oracle have failed to allow the report to be based on a PLSQL function returning the query.
  • Step1.Create a collection based on a query. The code for this should be placed in a before header process APEX_COLLECTION.CREATE_COLLECTION_FROM_QUERY(p_collection_name => ‘IR_TEST’,p_query => function_returning_query );
  • Step 2.Create an interactive report querying a collection. Select *From apex_collectionsWhere collection_name = ‘IR_TEST’;
cezarovidiu

Inventory Aging Query Oracle Applications R12 - 0 views

    • cezarovidiu
       
      SELECT ev1.INVENTORY_ITEM_ID,   ev1.ITEM_CODE ITEM,   XXRG_HR_PKG.get_org_name(:ORG_ID) ORGANIZATION,   ev1.DESCRIPTION,   SUM(Buk11) buk11,   SUM(Buk21) buk21 ,   SUM(Buk31) buk31 ,   SUM(Buk41) buk41,   SUM(Buk51) buk51  FROM   (SELECT ev1.INVENTORY_ITEM_ID,     ITEM_CODE,     DESCRIPTION,     (     CASE       WHEN age BETWEEN :BUK1_DAYS_FROM AND :BUK1_DAYS_TO       THEN SUM(aqty)       ELSE NULL     END) Buk11,     (     CASE       WHEN age BETWEEN :BUK2_DAYS_FROM AND :BUK2_DAYS_TO       THEN SUM(aqty)       ELSE NULL     END) Buk21,     (     CASE       WHEN age BETWEEN :BUK3_DAYS_FROM AND :BUK3_DAYS_TO       THEN SUM(aqty)       ELSE NULL     END) Buk31,     (     CASE       WHEN age BETWEEN :BUK4_DAYS_FROM AND :BUK4_DAYS_TO       THEN SUM(aqty)       ELSE NULL     END) Buk41,     (     CASE       WHEN age >= :BUK5_DAYS_FROM       THEN SUM(aqty)       ELSE NULL     END) Buk51   FROM     (SELECT        ITEM_CODE,       DESCRIPTION,       TRANSACTION_DATE,       TRANSACTION_QUANTITY,       SUM(TRANSACTION_QUANTITY) OVER(PARTITION BY INVENTORY_ITEM_ID ORDER BY TRANSACTION_ID,TRANSACTION_DATE)+ NVL(NQTY,0) BFF ,       (       CASE         WHEN TRANSACTION_QUANTITY > SUM(TRANSACTION_QUANTITY) OVER(PARTITION BY INVENTORY_ITEM_ID ORDER BY TRANSACTION_ID,TRANSACTION_DATE)+ NVL(NQTY,0)         THEN SUM(TRANSACTION_QUANTITY) OVER(PARTITION BY INVENTORY_ITEM_ID ORDER BY TRANSACTION_ID,TRANSACTION_DATE)                       +NVL(NQTY,0)         ELSE TRANSACTION_QUANTITY       END) AQTY       --,TCOST       ,       NVL(fnd_conc_date.string_to_date(:TILL_DATE),SYSDATE)-fnd_conc_date.string_to_date(TRANSACTION_DATE) AGE,       inventory_item_id     FROM       (SELECT V1.TRANSACTION_ID,         V1.ITEM_CODE,         V1.DESCRIPTION,         TRUNC(         CASE           WHEN V1.TRANSACTION_TYPE_ID = 4
cezarovidiu

You Probably Need Parallel Except When You Don't - 0 views

  • f you are running a large Oracle data warehouse you should be using parallel
  • Like all tools you have to use parallel correctly; no more would we think of using a wrench to hammer a nail then should you think parallel is the answer to all performance problems. Sometimes parallel will make things worse, sometimes parallel will make performance less predictable.
  • Parallel introduces additional work to a query, simplistically we need to: split the query into multiple parallel processes, execute them, wait for the processes to complete and finally coordinate the results. This all takes time to do. Our time saving comes from being able to process multiple smaller chunks of data simultaneously. If the time to execute the step in parallel is not significantly faster than doing it without parallel then the additional overhead may make parallel processing a slower option; this is typically the case with small tables where a full tablescan or an indexed access is fast. Use too few parallel processes and we will not gain much in performance, too many and we risk starving the database of resource for other work or even slow our own process as it waits for resource. If you have implemented some form of CPU resource management on your system you may find that you experience delays as your parallel slaves ‘wait their turn’
cezarovidiu

BI Tools, their SQL Generators, and Infobright - 1 views

  • The greatest benefit of columnar is to avoid disk I/O.  By choosing “select *”, you run the risk of losing that benefit. 
  • BI tools are here to stay, and they really help make visualization of analytics easy.  When working with Infobright, always take an extra second to review the generated queries.  The extra few seconds could mean seconds or minutes in saved query times.
  •  
    " The greatest benefit of columnar is to avoid disk I/O.  By choosing "select *", you run the risk of losing that benefit. "
cezarovidiu

InfiniDB - the high performance, column oriented analytic database - 0 views

  •  
    "Contributed by Calpont, InfiniDB Community Edition is an open source, scale-up analytics database engine for your data warehousing, business intelligence and read-intensive application needs. Enabled via MySQL® and purpose-built for an analytical workload with column-oriented technology at its core, the multi-threaded capabilities of InfiniDB Community Edition fully encompass query, transactional support and bulk load operations.  So come on in, grab a download and get started."
cezarovidiu

What is business intelligence (BI)? - Definition from WhatIs.com - 0 views

  • Business intelligence is a data analysis process aimed at boosting business performance by helping corporate executives and other end users make more informed decisions.
  • Business intelligence (BI) is a technology-driven process for analyzing data and presenting actionable information to help corporate executives, business managers and other end users make more informed business decisions.
  • BI encompasses a variety of tools, applications and methodologies that enable organizations to collect data from internal systems and external sources, prepare it for analysis, develop and run queries against the data, and create reports, dashboards and data visualizations to make the analytical results available to corporate decision makers as well as operational workers.
  • ...9 more annotations...
  • The potential benefits of business intelligence programs include accelerating and improving decision making; optimizing internal business processes; increasing operational efficiency; driving new revenues; and gaining competitive advantages over business rivals. BI systems can also help companies identify market trends and spot business problems that need to be addressed.
  • BI data can include historical information, as well as new data gathered from source systems as it is generated, enabling BI analysis to support both strategic and tactical decision-making processes.
  • BI programs can also incorporate forms of advanced analytics, such as data mining, predictive analytics, text mining, statistical analysis and big data analytics.
  • In many cases though, advanced analytics projects are conducted and managed by separate teams of data scientists, statisticians, predictive modelers and other skilled analytics professionals, while BI teams oversee more straightforward querying and analysis of business data.
  • Business intelligence data typically is stored in a data warehouse or smaller data marts that hold subsets of a company's information. In addition, Hadoop systems are increasingly being used within BI architectures as repositories or landing pads for BI and analytics data, especially for unstructured data, log files, sensor data and other types of big data. Before it's used in BI applications, raw data from different source systems must be integrated, consolidated and cleansed using data integration and data quality tools to ensure that users are analyzing accurate and consistent information.
  • In addition to BI managers, business intelligence teams generally include a mix of BI architects, BI developers, business analysts and data management professionals; business users often are also included to represent the business side and make sure its needs are met in the BI development process.
  • To help with that, a growing number of organizations are replacing traditional waterfall development with Agile BI and data warehousing approaches that use Agile software development techniques to break up BI projects into small chunks and deliver new functionality to end users on an incremental and iterative basis.
  • consultant Howard Dresner is credited with first proposing it in 1989 as an umbrella category for applying data analysis techniques to support business decision-making processes.
  • Business intelligence is sometimes used interchangeably with business analytics; in other cases, business analytics is used either more narrowly to refer to advanced data analytics or more broadly to include both BI and advanced analytics.
cezarovidiu

Entity Attribute Value in Magento - Magento tutorial lesson 19 - Magento - 0 views

  • Entity Attribute Value
  • number of attributes (properties and parameters) which can be used to describe them are potentially vast, but the number of attributes which will actually apply to a given entity are relatively modest.
  • sparse matrix
  • ...11 more annotations...
  • vertical database
  • open schema
  • more complex queries
  • Entity: The entity represents Magento data items
  • Attribute: The attributes represent data items that belong to an entity.
  • Value: The value is the simplest to understand as it is simply a value linked to an attribute
  • 1.1. What is Entity Attribute Value
  • 1.2. Entity table structure
  • At the very least, the attribute definitions table would contain the following columns: an attribute ID, attribute name, description, data type, and columns assisting input validation, e.g., maximum string length and regular expression, set of permissible values, etc.
  • The attribute or parameter:
  • The entity
cezarovidiu

Angel's BI Blog: Excel Data Explorer and the Twitter Search API - 0 views

  • its a very powerful query engine/data transformation tool.
  • what's fascinating is its ability to load web data from static html pages, tables on web pages, web services, etc. I had a need to search recent activity on Twitter and decided to test drive Data Explorer in Excel.
cezarovidiu

Magic Quadrant for Advanced Analytics Platforms - 1 views

  • Gartner defines advanced analytics as, "the analysis of all kinds of data using sophisticated quantitative methods (for example, statistics, descriptive and predictive data mining, simulation and optimization) to produce insights that traditional approaches to business intelligence (BI) — such as query and reporting — are unlikely to discover."
  • packaged analytics applications that target specific business domains
  • Revolution Analytics
cezarovidiu

Hadoop HBase 1.0 debuts amid stiff NoSQL competition | InfoWorld - 0 views

  • Databases consisting of billions of rows and columns can be stored in HBase and retrieved via conventional SQL queries, and an HBase database can scale out by simply adding nodes to an existing cluster.
1 - 20 of 22 Next ›
Showing 20 items per page