Skip to main content

Home/ BI-TAGS/ Group items tagged map

Rss Feed Group items tagged

cezarovidiu

Choose Your Own Adventure: Map Buyer Personas in a Content Marketing Program | ClickZ - 0 views

  • learn more about the brand or to be entertained or educated in some way.
  • Mapping out the ifs-ands-or-buts pathways for each persona can become complicated, so automation solutions such as Infusionsoft that offer a drag-and-drop interface can make the mapping of the funnel stages manageable.
cezarovidiu

Rittman Mead Consulting - The Changing World of Business Intelligence - 0 views

  • Schema on write This is the traditional approach for Business Intelligence. A model, often dimensional, is built as part of the design process. This model is an abstraction of the complexity of the underlying systems, put in business terms. The purpose of the model is to allow the business users to interrogate the data in a way they understand.
  • The model is instantiated through physical database tables and the date is loaded through an ETL (extract, transform and load) process that takes data from one or more source systems and transforms it to fit the model, then loads it into the model.
  • The key thing is that the model is determined before the data is finally written and the users are very much guided or driven by the model in how they query the data and what results they can get from the system. The designer must anticipate the queries and requests in advance of the user asking the questions.
  • ...3 more annotations...
  • Schema on read Schema on read works on a different principle and is more common in the Big Data world. The data is not transformed in any way when it is stored, the data store acts as a big bucket. The modelling of the data only occurs when the data is read. Map/Reduce is the clearest example, the mapping is the understanding of the data structure. Hadoop is a large distributed file system, which is very good at storing large volumes of data, this is potential. It is only the mapping of this data that provides value, this is done when the data is read, not written.
  • New World Order So whereas Business Intelligence used to always be driven by the model, the ETL process to populate the model and the reporting tool to query the model, there is now an approach where the data is collected its raw form, and advanced statistical or analytical tools are used to interrogate the data. An example of one such tool is R.
  • The driver for which approach to use is often driven by what the user wants to find out. If the question is clearly formed and the sources of data that are required to answer it well understood, for example how many units of a product have we sold, then the traditional schema on write approach is best.
cezarovidiu

Tips and tricks for creating reports - Microsoft Power BI Preview - 0 views

  • When you create a Map, you're often looking to plot Countries, States, and Cities.  If you use name columns after the geographic designation it will help Bing guess what you're looking to display. For example, if you have a field of US state names such as "California" and "Washington", Bing might return the location of Washington, DC instead of Washington State for the word "Washington".  Naming the column "State" will improve the geocoding.  The same is goes for columns named "Country", "State", and "City".  
cezarovidiu

Magic Quadrant for Business Intelligence and Analytics Platforms - 0 views

  • Integration BI infrastructure: All tools in the platform use the same security, metadata, administration, portal integration, object model and query engine, and should share the same look and feel. Metadata management: Tools should leverage the same metadata, and the tools should provide a robust way to search, capture, store, reuse and publish metadata objects, such as dimensions, hierarchies, measures, performance metrics and report layout objects. Development tools: The platform should provide a set of programmatic and visual tools, coupled with a software developer's kit for creating analytic applications, integrating them into a business process, and/or embedding them in another application. Collaboration: Enables users to share and discuss information and analytic content, and/or to manage hierarchies and metrics via discussion threads, chat and annotations.
  • Information Delivery Reporting: Provides the ability to create formatted and interactive reports, with or without parameters, with highly scalable distribution and scheduling capabilities. Dashboards: Includes the ability to publish Web-based or mobile reports with intuitive interactive displays that indicate the state of a performance metric compared with a goal or target value. Increasingly, dashboards are used to disseminate real-time data from operational applications, or in conjunction with a complex-event processing engine. Ad hoc query: Enables users to ask their own questions of the data, without relying on IT to create a report. In particular, the tools must have a robust semantic layer to enable users to navigate available data sources. Microsoft Office integration: Sometimes, Microsoft Office (particularly Excel) acts as the reporting or analytics client. In these cases, it is vital that the tool provides integration with Microsoft Office, including support for document and presentation formats, formulas, data "refreshes" and pivot tables. Advanced integration includes cell locking and write-back. Search-based BI: Applies a search index to structured and unstructured data sources and maps them into a classification structure of dimensions and measures that users can easily navigate and explore using a search interface. Mobile BI: Enables organizations to deliver analytic content to mobile devices in a publishing and/or interactive mode, and takes advantage of the mobile client's location awareness.
  • Analysis Online analytical processing (OLAP): Enables users to analyze data with fast query and calculation performance, enabling a style of analysis known as "slicing and dicing." Users are able to navigate multidimensional drill paths. They also have the ability to write back values to a proprietary database for planning and "what if" modeling purposes. This capability could span a variety of data architectures (such as relational or multidimensional) and storage architectures (such as disk-based or in-memory). Interactive visualization: Gives users the ability to display numerous aspects of the data more efficiently by using interactive pictures and charts, instead of rows and columns. Predictive modeling and data mining: Enables organizations to classify categorical variables, and to estimate continuous variables using mathematical algorithms. Scorecards: These take the metrics displayed in a dashboard a step further by applying them to a strategy map that aligns key performance indicators (KPIs) with a strategic objective. Prescriptive modeling, simulation and optimization: Supports decision making by enabling organizations to select the correct value of a variable based on a set of constraints for deterministic processes, and by modeling outcomes for stochastic processes.
  • ...7 more annotations...
  • These capabilities enable organizations to build precise systems of classification and measurement to support decision making and improve performance. BI and analytic platforms enable companies to measure and improve the metrics that matter most to their businesses, such as sales, profits, costs, quality defects, safety incidents, customer satisfaction, on-time delivery and so on. BI and analytic platforms also enable organizations to classify the dimensions of their businesses — such as their customers, products and employees — with more granular precision. With these capabilities, marketers can better understand which customers are most likely to churn. HR managers can better understand which attributes to look for when recruiting top performers. Supply chain managers can better understand which inventory allocation levels will keep costs low without increasing out-of-stock incidents.
  • descriptive, diagnostic, predictive and prescriptive analytics
  • "descriptive"
  • diagnostic
  • data discovery vendors — such as QlikTech, Salient Management Company, Tableau Software and Tibco Spotfire — received more positive feedback than vendors offering OLAP cube and semantic-layer-based architectures.
  • Microsoft Excel users are often disaffected business BI users who are unable to conduct the analysis they want using enterprise, IT-centric tools. Since these users are the typical target users of data discovery tool vendors, Microsoft's aggressive plans to enhance Excel will likely pose an additional competitive threat beyond the mainstreaming and integration of data discovery features as part of the other leading, IT-centric enterprise platforms.
  • Building on the in-memory capabilities of PowerPivot in SQL Server 2012, Microsoft introduced a fully in-memory version of Microsoft Analysis Services cubes, based on the same data structure as PowerPivot, to address the needs of organizations that are turning to newer in-memory OLAP architectures over traditional, multidimensional OLAP architectures to support dynamic and interactive analysis of large datasets. Above-average performance ratings suggest that customers are happy with the in-memory improvements in SQL Server 2012 compared with SQL Server 2008 R2, which ranks below the survey average.
  •  
    "Gartner defines the business intelligence (BI) and analytics platform market as a software platform that delivers 15 capabilities across three categories: integration, information delivery and analysis."
cezarovidiu

Installing Hadoop for Fedora & Oracle Linux(Single Node Cluster) | accretion infinity - 0 views

  • Hadoop is a framework written in Java for running applications on large clusters of commodity hardware and incorporates features similar to those of the Google File System (GFS) and of the Map Reduce computing paradigm. Hadoop’s HDFS is a highly fault-tolerant distributed file system and, like Hadoop in general, designed to be deployed on low-cost hardware. It provides high throughput access to application data and is suitable for applications that have large data sets.
  • Some of the Hadoop projects we will talk about are: HDFS : A distributed filesystem that runs on large clusters of commodity machines. Map Reduce: A distributed data processing model and execution environment that runs on large clusters of commodity machines. Pig: A data flow language and execution environment for exploring very large datasets. Pig runs on HDFS and MapReduce clusters. HBase: A distributed, column-oriented database. HBase uses HDFS for its underlying storage, and supports both batch-style computations using MapReduce and point queries (random reads). ZooKeeper: A distributed, highly available coordination service. ZooKeeper provides primitives such as distributed locks that can be used for building distributed applications. Oozie: Oozie is a workflow scheduler system to manage Apache Hadoop jobs.
  • Oracle Linux as the operating system and Hadoop 1.1.2 or 1.2.0
cezarovidiu

BI Brief - Four Legs of a Successful Business Intelligence (BI) Project Team - 0 views

  • 1. Project Sponsorship and Governance 2. Project Management 3. Development Team (Core Team) 4. Extended Project Team
  • 1. Project Sponsorship and Governance IT and the business should form a BI steering committee to sponsor and govern design, development, deployment, and ongoing support. It needs both the CIO and a business executive, such as CFO, COO, or a senior VP of marketing/sales to commit budget, time, and resources. The business sponsor needs the project to succeed. The CIO is committed to what is being built and how.
  • 2. Project Management Project management includes managing daily tasks, reporting status, and communicating to the extended project team, steering committee, and affected business users. The project management team needs extensive business knowledge, BI expertise, DW architecture background, and people management, project management, and communications skills. The project management team includes three functions or members: Project development manager - Responsible for deliverables, managing team resources, monitoring tasks, reporting status, and communications. Requires a hands-on IT manager with a background in iterative development. Must understand the changes caused by this approach and the impact on the business, project resources, schedule and the trade-offs. Business advisor - Works within the sponsoring business organization. Responsible for the deliverables of the business resources on the project's extended team. Serves as the business advocate on the project team and the project advocate within the business community. Often, the business advocate is a project co-manager who defers to the IT project manager the daily IT tasks but oversees the budget and business deliverables. BI/DW project advisor - Has enough expertise with architectures and technologies to guides the project team on their use. Ensures that architecture, data models, databases, ETL code, and BI tools are all being used effectively and conform to best practices and standards.
  • ...2 more annotations...
  • 3. Development Team (Core Team) The core project team is divided into four sub-teams: Business requirements - This sub-team may have business people who understand IT systems, or IT people who understand the business. In either case, the team represents the business and their interests. They are responsible for gathering and prioritizing business needs; translating them into IT systems requirements; interacting with the business on the data quality and completeness; and ensuring the business provides feedback on how well the solutions generated meet their needs. BI architecture - Develops the overall BI architecture, selects the appropriate technology, creates the data models, maps the overall data workflow from source systems to BI analytics, and oversees the ETL and BI development teams from a technical perspective. ETL development - Receives the business and data requirements, as well as the target data models to be used by BI analytics. Develops the ETL code needed to gather data from the appropriate source systems into the BI databases. Often, a system analyst who is a expert in the source systems such as SAP is part of the team to provide knowledge of the data sources, customizations, and data quality. BI development - Create the reports or analytics that the business users will interact with to do their jobs. This is often a very iterative process and requires much interaction with the business users.
  • 4. Extended Project Team There are several functions required by the project team that are often accomplished through an "extended" team: Players - A group of business users are signed up to "play with" or test the BI analytics and reports as they are developed to provide feedback to the core development team. This is a virtual team that gets together at specific periods of the project but they are committed to this role during those periods. Testers - A group of resources are gathered, similarly to the virtual team above, to perform more extensive QA testing of the BI analytics, ETL processes, and overall systems testing. You may have project members test other members' work, such as the ETL team test the BI analytics and visa versa. Operators - IT operations is often separated from the development team but it is critical that they are involved from the beginning of the project to ensure that the systems are developed and deployed within your company's infrastructure. Key functions are database administration, systems administration, and networks. In addition, this extended team may also include help desk and training resources if they are usually provided outside of development.
cezarovidiu

What Skills Does an Oracle BI Developer Need in 2011? - 0 views

  • OBIEE 11g skills, both in terms of new functionality (mapping, analyses, KPIs and Scorecards etc) and new infrastructure (WebLogic, EM, OPSS etc) A smattering of Essbase skills, focused mainly on the integration with OBIEE and Essbase (and the many workarounds and gotchas) Good ODI skills, both in terms of the basics, but also being able to write knowledge modules, integrate with OBIEE, deployment and migration Solid database skills – OBIEE gave the illusion through aggregates etc that database tuning was redundant, but time has shown it’s by far the biggest success factor in a project – get the database design and optimisation wrong, and your project is toast. You need to know partitioning, materialized views, index types, and increasingly, you need to get yourself on an Exadata project as customers are buying the technology but you can’t teach it to yourself at home BI Apps skills, but watch out for everything changing when BI Apps 11g comes out, and be prepared to learn the Fusion Apps and JDeveloper if you want to stay in the game Looking to the future, keep an eye on technologies such as in-memory (TimesTen), mid-tier caching (Coherence), plus technologies such as Business Activity Monitoring (BAM), “big data” (Hadoop, large data sets, NoSQL), complex event processing and maybe products such as Qlikview, just in case Oracle buys them, or at least to know what the competition are up to, or more importantly pitching to your boss
  • The other thing to bear in mind of course, if you’re an Oracle BI developer, is that you need to have great business, communication and data modeling skills.
cezarovidiu

Filling a Critical Role in Business Today: The Data Translator - Microsoft Business Int... - 0 views

  • a lot of articles calling data scientists and statisticians the jobs of the future
  • there are more immediate needs that, when addressed, will have a much greater business impact.
  • Right now we have huge opportunities to make the data more accessible, more “joinable” and more consumable. Leaders don’t want more data – they want more information they can use to run their businesses.
  • ...5 more annotations...
  • Every company has hundreds of millions of records about their sales, expenses, employees and so on, with dozens of insights yet to be discovered through simple comparison or triangulation of relevant data.
  • Why don’t we focus on this? I think because it’s very difficult to do – being successful in this “data translator” role requires a unique set of skills and knowledge, the combination of which I call the BASE skillset: Business understanding Ability to synthetize and simplify Storytelling skills Expertise in data visualization
  • Business Understanding This one seems obvious, but it doesn’t mean simply understanding the financials of a business. Rather, it means truly knowing the operational details, the incentives, the install base, market growth, penetration, the competition, etc. An analyst can’t just know the technical aspect of a report or the math behind the numbers, but what is truly driving a pattern in terms of product quality, competition, incentives and/or offerings. The best analysts are able to mathematically isolate the key levers of a trend and then suggest actions to react to or take advantage of those trends. Ability to Synthetize and Simplify This is, in my opinion, the most underrated and underappreciated skill. Combing through thousands of data points and netting out 3-4 key issues in under 10 minutes, and then communicating these to a group of execs with very different analytical skills, is truly difficult. The key is to make it simple but not simplistic, which means you still capture the complexity even as you get to the few core insights. It requires a very thorough effort to gather all the relevant information before categorizing, prioritizing and deciding if it is significant. After a while, you become an expert and can sniff things out quickly. At the same time, there is the danger of missing anomalies when you jump to conclusions based only on a summary look.
  • Storytelling Skills There are stages that should be followed when explaining complex ideas, something data translators are frequently expected to do. The best storytellers start by giving context and trying to couple the current discussion to something the audience already knows, ensuring the story is well structured and connected. We have to move from a “buffet style” business review with thousands of numbers packed in tables to a layered approach that will guide the audience to focus first on the most relevant messages, diving deeper only when necessary. Minto Pyramid Principles, which are built around a process for organizing thought and communication, are helpful in making sure you really focus on what is important and relevant, versus being obsessed in telling every fact. Expertise in Data Visualization I am glad to finally see so much focus on Information Visualization and I believe this is correlated to the explosion of data. Traditional methods of organizing data do not facilitate an intuitive understanding of key information points or trends. For instance, the two examples below contain data on car sales across the U.S. The first, an alphabetized list, is much less intuitive than the second, which shows those sales on a map in Power View. With Power View, right away you can identify the states with the highest sales: CA, FL, TX, NY. (Workbook available here)
  • There is no better way to see patterns or trends than data visualization, making expertise in this area – both technical and analytical – critical for data translators.
cezarovidiu

How to join GL tables with XLA (SubLedger Accounting) tables - 0 views

  • GL_JE_BATCHES (je_batch_id)                                   => GL_JE_HEADERS (je_batch_id) GL_JE_HEADERS (je_header_id)                                  => GL_JE_LINES (je_header_id) GL_JE_LINES (je_header_id,  je_line_num)                      => GL_IMPORT_REFERENCES (je_header_id, je_line_num) GL_IMPORT_REFERENCES (gl_sl_link_table, gl_sl_link_id)        => XLA_AE_LINES (gl_sl_link_table, gl_sl_link_id) XLA_AE_LINES (application_id, ae_header_id)                   => XLA_AE_HEADERS (application_id, ae_header_id)  XLA_AE_HEADERS (application_id, event_id)                     => XLA_EVENTS (application_id, event_id)    XLA_EVENTS (application_id, entity_id)                        => XLA.XLA_TRANSACTION_ENTITIES (application_id, entity_id) 
  • The source_id_int_1 column of xla.xla_transaction_entities stores the primary_id value for the transactions. You can join the xla.xla_transaction_entities table with the corresponding transactions table for obtaining additional information of the transaction. For e.g you join the xla.xla_transaction_entities table with ra_customer_trx_all for obtaining receivables transactions information or with mtl_material_transactions table for obtaining material transactions information. The entity_id mappings can be obtained from the XLA_ENTITY_ID_MAPPINGS table
  •  
    "How to join GL tables with XLA (SubLedger Accounting) tables"
cezarovidiu

Focus on Valuable Data - Not Big Data - to Boost Conversions and ROI | ClickZ - 0 views

  • Big Data has been all the rage. But fast data, even if it is small, can be more valuable than complicated masses of information.
  • Here's why: All the focus on "bigger is better" has overlooked the fact that most Big Data segments have not been validated with a business application or value.
  • Those kinds of analytics can help you find the right streams to access and work with, and also can help you build out robust programs that identify valuable customers.
  • ...1 more annotation...
  • 1) Your First-Party Data: The primary and most valuable data set you can access, first-party data encompasses transactional and other customer-level profile information you have on your customers. It could also include your own off-line segmentation analysis that allows you to map a customer to a customer profile around which you build your marketing programs. This can also include your analytics or other on-site tracking data, which can deliver behavioral insight to your consumers. This data can be difficult to export from its current environment due to the ad hoc nature of the data, but, if possible, look at ways to make this information accessible to your digital sites. 2) Third-Party Data: A consumer's broader Web browsing and buying history can now be accessed in session to provide you with more context on their likes and habits. Data management platforms (DMPs) and other data aggregators are accelerating this offering and, just as importantly, the availability of this type of data. This is invaluable in the context of new visitors who you know nothing about historically. 3) Real-Time Behavior: Let's not forget what our customers are telling us with each click. We get enamored with our predictive modeling to the point that we do not see the tell-tale signs as they are happening. Take the time to stop, look, and react. Your analytic tools, personalization tools, and other software-as-a-service (SaaS) platforms can help you trigger alternate site experiences based on every click you see.
1 - 18 of 18
Showing 20 items per page