Skip to main content

Home/ BI-TAGS/ Group items tagged data

Rss Feed Group items tagged

cezarovidiu

What is business intelligence (BI)? - Definition from WhatIs.com - 0 views

  • Business intelligence is a data analysis process aimed at boosting business performance by helping corporate executives and other end users make more informed decisions.
  • Business intelligence (BI) is a technology-driven process for analyzing data and presenting actionable information to help corporate executives, business managers and other end users make more informed business decisions.
  • BI encompasses a variety of tools, applications and methodologies that enable organizations to collect data from internal systems and external sources, prepare it for analysis, develop and run queries against the data, and create reports, dashboards and data visualizations to make the analytical results available to corporate decision makers as well as operational workers.
  • ...9 more annotations...
  • The potential benefits of business intelligence programs include accelerating and improving decision making; optimizing internal business processes; increasing operational efficiency; driving new revenues; and gaining competitive advantages over business rivals. BI systems can also help companies identify market trends and spot business problems that need to be addressed.
  • BI data can include historical information, as well as new data gathered from source systems as it is generated, enabling BI analysis to support both strategic and tactical decision-making processes.
  • BI programs can also incorporate forms of advanced analytics, such as data mining, predictive analytics, text mining, statistical analysis and big data analytics.
  • In many cases though, advanced analytics projects are conducted and managed by separate teams of data scientists, statisticians, predictive modelers and other skilled analytics professionals, while BI teams oversee more straightforward querying and analysis of business data.
  • Business intelligence data typically is stored in a data warehouse or smaller data marts that hold subsets of a company's information. In addition, Hadoop systems are increasingly being used within BI architectures as repositories or landing pads for BI and analytics data, especially for unstructured data, log files, sensor data and other types of big data. Before it's used in BI applications, raw data from different source systems must be integrated, consolidated and cleansed using data integration and data quality tools to ensure that users are analyzing accurate and consistent information.
  • In addition to BI managers, business intelligence teams generally include a mix of BI architects, BI developers, business analysts and data management professionals; business users often are also included to represent the business side and make sure its needs are met in the BI development process.
  • To help with that, a growing number of organizations are replacing traditional waterfall development with Agile BI and data warehousing approaches that use Agile software development techniques to break up BI projects into small chunks and deliver new functionality to end users on an incremental and iterative basis.
  • consultant Howard Dresner is credited with first proposing it in 1989 as an umbrella category for applying data analysis techniques to support business decision-making processes.
  • Business intelligence is sometimes used interchangeably with business analytics; in other cases, business analytics is used either more narrowly to refer to advanced data analytics or more broadly to include both BI and advanced analytics.
cezarovidiu

Data Quality, Data Governance, and Master Data Management (MDM) - 0 views

  • Modern business applications produce ever more relevant and actionable information for decision makers, but in many cases the data sources are fragmented and inconsistent. Despite tremendous advancements at the application layer, nearly all IT initiatives succeed or fail based on the quality and consistency of the underlying data.
  • CIOs are responsible for making information available to their businesses in a consistent and timely basis, but in most organizations, information management is seen as a delegated set of tasks and is not the CIO’s top priority.
  • “Key initiatives such as master data management, data virtualization, data quality, data integration and data governance are employed by just a fraction of organizations that should be mastering the science of information management,”
  • ...1 more annotation...
  • While CIOs are aware that effective information management results in faster decision-making, according to the Ventana study, only 43% of organizations have undertaken information management initiatives in data governance, data integration, data quality, master data management and data virtualization during the last two years, and less than one fifth have completed those projects. The largest obstacles to completing information management projects are insufficient staffing (68%), inadequate budget (63%) and insufficient training and skills (59%).
  •  
    "Modern business applications produce ever more relevant and actionable information for decision makers, but in many cases the data sources are fragmented and inconsistent. Despite tremendous advancements at the application layer, nearly all IT initiatives succeed or fail based on the quality and consistency of the underlying data."
cezarovidiu

Focus on Valuable Data - Not Big Data - to Boost Conversions and ROI | ClickZ - 0 views

  • Big Data has been all the rage. But fast data, even if it is small, can be more valuable than complicated masses of information.
  • Here's why: All the focus on "bigger is better" has overlooked the fact that most Big Data segments have not been validated with a business application or value.
  • Those kinds of analytics can help you find the right streams to access and work with, and also can help you build out robust programs that identify valuable customers.
  • ...1 more annotation...
  • 1) Your First-Party Data: The primary and most valuable data set you can access, first-party data encompasses transactional and other customer-level profile information you have on your customers. It could also include your own off-line segmentation analysis that allows you to map a customer to a customer profile around which you build your marketing programs. This can also include your analytics or other on-site tracking data, which can deliver behavioral insight to your consumers. This data can be difficult to export from its current environment due to the ad hoc nature of the data, but, if possible, look at ways to make this information accessible to your digital sites. 2) Third-Party Data: A consumer's broader Web browsing and buying history can now be accessed in session to provide you with more context on their likes and habits. Data management platforms (DMPs) and other data aggregators are accelerating this offering and, just as importantly, the availability of this type of data. This is invaluable in the context of new visitors who you know nothing about historically. 3) Real-Time Behavior: Let's not forget what our customers are telling us with each click. We get enamored with our predictive modeling to the point that we do not see the tell-tale signs as they are happening. Take the time to stop, look, and react. Your analytic tools, personalization tools, and other software-as-a-service (SaaS) platforms can help you trigger alternate site experiences based on every click you see.
cezarovidiu

Rittman Mead Consulting - The Changing World of Business Intelligence - 0 views

  • Schema on write This is the traditional approach for Business Intelligence. A model, often dimensional, is built as part of the design process. This model is an abstraction of the complexity of the underlying systems, put in business terms. The purpose of the model is to allow the business users to interrogate the data in a way they understand.
  • The model is instantiated through physical database tables and the date is loaded through an ETL (extract, transform and load) process that takes data from one or more source systems and transforms it to fit the model, then loads it into the model.
  • The key thing is that the model is determined before the data is finally written and the users are very much guided or driven by the model in how they query the data and what results they can get from the system. The designer must anticipate the queries and requests in advance of the user asking the questions.
  • ...3 more annotations...
  • Schema on read Schema on read works on a different principle and is more common in the Big Data world. The data is not transformed in any way when it is stored, the data store acts as a big bucket. The modelling of the data only occurs when the data is read. Map/Reduce is the clearest example, the mapping is the understanding of the data structure. Hadoop is a large distributed file system, which is very good at storing large volumes of data, this is potential. It is only the mapping of this data that provides value, this is done when the data is read, not written.
  • New World Order So whereas Business Intelligence used to always be driven by the model, the ETL process to populate the model and the reporting tool to query the model, there is now an approach where the data is collected its raw form, and advanced statistical or analytical tools are used to interrogate the data. An example of one such tool is R.
  • The driver for which approach to use is often driven by what the user wants to find out. If the question is clearly formed and the sources of data that are required to answer it well understood, for example how many units of a product have we sold, then the traditional schema on write approach is best.
cezarovidiu

Filling a Critical Role in Business Today: The Data Translator - Microsoft Business Int... - 0 views

  • a lot of articles calling data scientists and statisticians the jobs of the future
  • there are more immediate needs that, when addressed, will have a much greater business impact.
  • Right now we have huge opportunities to make the data more accessible, more “joinable” and more consumable. Leaders don’t want more data – they want more information they can use to run their businesses.
  • ...5 more annotations...
  • Every company has hundreds of millions of records about their sales, expenses, employees and so on, with dozens of insights yet to be discovered through simple comparison or triangulation of relevant data.
  • Why don’t we focus on this? I think because it’s very difficult to do – being successful in this “data translator” role requires a unique set of skills and knowledge, the combination of which I call the BASE skillset: Business understanding Ability to synthetize and simplify Storytelling skills Expertise in data visualization
  • Business Understanding This one seems obvious, but it doesn’t mean simply understanding the financials of a business. Rather, it means truly knowing the operational details, the incentives, the install base, market growth, penetration, the competition, etc. An analyst can’t just know the technical aspect of a report or the math behind the numbers, but what is truly driving a pattern in terms of product quality, competition, incentives and/or offerings. The best analysts are able to mathematically isolate the key levers of a trend and then suggest actions to react to or take advantage of those trends. Ability to Synthetize and Simplify This is, in my opinion, the most underrated and underappreciated skill. Combing through thousands of data points and netting out 3-4 key issues in under 10 minutes, and then communicating these to a group of execs with very different analytical skills, is truly difficult. The key is to make it simple but not simplistic, which means you still capture the complexity even as you get to the few core insights. It requires a very thorough effort to gather all the relevant information before categorizing, prioritizing and deciding if it is significant. After a while, you become an expert and can sniff things out quickly. At the same time, there is the danger of missing anomalies when you jump to conclusions based only on a summary look.
  • Storytelling Skills There are stages that should be followed when explaining complex ideas, something data translators are frequently expected to do. The best storytellers start by giving context and trying to couple the current discussion to something the audience already knows, ensuring the story is well structured and connected. We have to move from a “buffet style” business review with thousands of numbers packed in tables to a layered approach that will guide the audience to focus first on the most relevant messages, diving deeper only when necessary. Minto Pyramid Principles, which are built around a process for organizing thought and communication, are helpful in making sure you really focus on what is important and relevant, versus being obsessed in telling every fact. Expertise in Data Visualization I am glad to finally see so much focus on Information Visualization and I believe this is correlated to the explosion of data. Traditional methods of organizing data do not facilitate an intuitive understanding of key information points or trends. For instance, the two examples below contain data on car sales across the U.S. The first, an alphabetized list, is much less intuitive than the second, which shows those sales on a map in Power View. With Power View, right away you can identify the states with the highest sales: CA, FL, TX, NY. (Workbook available here)
  • There is no better way to see patterns or trends than data visualization, making expertise in this area – both technical and analytical – critical for data translators.
cezarovidiu

Gartner Positions Oracle in Leaders Quadrant for Master Data Management of Product Data... - 0 views

  • For the fourth consecutive year, Gartner, Inc. has named Oracle as a Leader in its “Magic Quadrant for Master Data Management of Product Data Solutions.” (1)
  • “MDM is a technology-enabled discipline in which business and IT staff work together to ensure the uniformity, accuracy, stewardship, semantic consistency and accountability of the enterprise's official, shared master data assets. Master data is the consistent and uniform set of identifiers and extended attributes that describes the core entities of the enterprise, such as customers, prospects, citizens, suppliers, sites, hierarchies and chart of accounts,” according to Gartner.
  • By enabling organizations to consolidate product information from heterogeneous systems, Oracle Product Hub creates a single view of product information that can be leveraged and shared across functional departments in the enterprise, as well as externally with trading partners.
  • ...1 more annotation...
  • "In any product company, accurate product information is a foundation for all major business initiatives, and this requires a robust, comprehensive and flexible product MDM solution,” said Jon Chorley, vice president, supply chain management product strategy, Oracle. "We believe Oracle's position in Gartner's Magic Quadrant for Master Data Management of Product Data Solutions highlights our ability to provide best-in-class functionality across the industry’s most complete MDM portfolio. By using Oracle MDM solutions, companies can obtain a high-quality, common enterprise product record and are better able to support their key business initiatives.”
  •  
    "Gartner Positions Oracle in Leaders Quadrant for Master Data Management of Product Data Solutions"
cezarovidiu

Magic Quadrant for Business Intelligence and Analytics Platforms - 0 views

  • Integration BI infrastructure: All tools in the platform use the same security, metadata, administration, portal integration, object model and query engine, and should share the same look and feel. Metadata management: Tools should leverage the same metadata, and the tools should provide a robust way to search, capture, store, reuse and publish metadata objects, such as dimensions, hierarchies, measures, performance metrics and report layout objects. Development tools: The platform should provide a set of programmatic and visual tools, coupled with a software developer's kit for creating analytic applications, integrating them into a business process, and/or embedding them in another application. Collaboration: Enables users to share and discuss information and analytic content, and/or to manage hierarchies and metrics via discussion threads, chat and annotations.
  • Information Delivery Reporting: Provides the ability to create formatted and interactive reports, with or without parameters, with highly scalable distribution and scheduling capabilities. Dashboards: Includes the ability to publish Web-based or mobile reports with intuitive interactive displays that indicate the state of a performance metric compared with a goal or target value. Increasingly, dashboards are used to disseminate real-time data from operational applications, or in conjunction with a complex-event processing engine. Ad hoc query: Enables users to ask their own questions of the data, without relying on IT to create a report. In particular, the tools must have a robust semantic layer to enable users to navigate available data sources. Microsoft Office integration: Sometimes, Microsoft Office (particularly Excel) acts as the reporting or analytics client. In these cases, it is vital that the tool provides integration with Microsoft Office, including support for document and presentation formats, formulas, data "refreshes" and pivot tables. Advanced integration includes cell locking and write-back. Search-based BI: Applies a search index to structured and unstructured data sources and maps them into a classification structure of dimensions and measures that users can easily navigate and explore using a search interface. Mobile BI: Enables organizations to deliver analytic content to mobile devices in a publishing and/or interactive mode, and takes advantage of the mobile client's location awareness.
  • Analysis Online analytical processing (OLAP): Enables users to analyze data with fast query and calculation performance, enabling a style of analysis known as "slicing and dicing." Users are able to navigate multidimensional drill paths. They also have the ability to write back values to a proprietary database for planning and "what if" modeling purposes. This capability could span a variety of data architectures (such as relational or multidimensional) and storage architectures (such as disk-based or in-memory). Interactive visualization: Gives users the ability to display numerous aspects of the data more efficiently by using interactive pictures and charts, instead of rows and columns. Predictive modeling and data mining: Enables organizations to classify categorical variables, and to estimate continuous variables using mathematical algorithms. Scorecards: These take the metrics displayed in a dashboard a step further by applying them to a strategy map that aligns key performance indicators (KPIs) with a strategic objective. Prescriptive modeling, simulation and optimization: Supports decision making by enabling organizations to select the correct value of a variable based on a set of constraints for deterministic processes, and by modeling outcomes for stochastic processes.
  • ...7 more annotations...
  • These capabilities enable organizations to build precise systems of classification and measurement to support decision making and improve performance. BI and analytic platforms enable companies to measure and improve the metrics that matter most to their businesses, such as sales, profits, costs, quality defects, safety incidents, customer satisfaction, on-time delivery and so on. BI and analytic platforms also enable organizations to classify the dimensions of their businesses — such as their customers, products and employees — with more granular precision. With these capabilities, marketers can better understand which customers are most likely to churn. HR managers can better understand which attributes to look for when recruiting top performers. Supply chain managers can better understand which inventory allocation levels will keep costs low without increasing out-of-stock incidents.
  • descriptive, diagnostic, predictive and prescriptive analytics
  • "descriptive"
  • diagnostic
  • data discovery vendors — such as QlikTech, Salient Management Company, Tableau Software and Tibco Spotfire — received more positive feedback than vendors offering OLAP cube and semantic-layer-based architectures.
  • Microsoft Excel users are often disaffected business BI users who are unable to conduct the analysis they want using enterprise, IT-centric tools. Since these users are the typical target users of data discovery tool vendors, Microsoft's aggressive plans to enhance Excel will likely pose an additional competitive threat beyond the mainstreaming and integration of data discovery features as part of the other leading, IT-centric enterprise platforms.
  • Building on the in-memory capabilities of PowerPivot in SQL Server 2012, Microsoft introduced a fully in-memory version of Microsoft Analysis Services cubes, based on the same data structure as PowerPivot, to address the needs of organizations that are turning to newer in-memory OLAP architectures over traditional, multidimensional OLAP architectures to support dynamic and interactive analysis of large datasets. Above-average performance ratings suggest that customers are happy with the in-memory improvements in SQL Server 2012 compared with SQL Server 2008 R2, which ranks below the survey average.
  •  
    "Gartner defines the business intelligence (BI) and analytics platform market as a software platform that delivers 15 capabilities across three categories: integration, information delivery and analysis."
cezarovidiu

Big data: The next frontier for innovation, competition, and productivity | McKinsey & ... - 0 views

  • The amount of data in our world has been exploding, and analyzing large data sets—so-called big data—will become a key basis of competition, underpinning new waves of productivity growth, innovation, and consumer surplus, according to research by MGI and McKinsey's Business Technology Office.
  • For example, a retailer using big data to the full could increase its operating margin by more than 60 percent.
  • important factor of production, alongside labor and capital.
  • ...9 more annotations...
  • five broad ways in which using big data can create value
  • Leading companies are using data collection and analysis to conduct controlled experiments to make better management decisions
  • others are using data for basic low-frequency forecasting to high-frequency nowcasting to adjust their business levers just in time.
  • big data allows ever-narrower segmentation of customers and therefore much more precisely tailored products or services.
  • Fourth, sophisticated analytics can substantially improve decision-making
  • big data can be used to improve the development of the next generation of products and services.
  • The use of big data will become a key basis of competition and growth for individual firms.
  • For example, we estimate that a retailer using big data to the full has the potential to increase its operating margin by more than 60 percent.
  • The computer and electronic products and information sectors, as well as finance and insurance, and government are poised to gain substantially from the use of big data.
cezarovidiu

Why BI projects fail -- and how to succeed instead | InfoWorld - 0 views

  • A successful initiative starts with a good strategy, and a good strategy starts with identifying the business need.
  • The balanced scorecard is one popular methodology for linking strategy, technology, and performance management. Other methodologies, such as applied information economics, combine statistical analysis, portfolio theory, and decision science in order to help firms calculate the economic value of better information. Whether you use a published methodology or develop your own approach in-house, the important point is to make sure your BI activities are keyed to generating real business value, not merely creating pretty, but useless, dashboards and reports.
  • Next, ask: What data do we wish we had and how would that lead to different decisions? The answers to these questions form top-level requirements for any BI project.
  • ...10 more annotations...
  • Instead a team of data experts, data analysts, and business experts must come together with the right technical expertise. This usually means bringing in outside help, though that help needs to be able to talk to management and talk tech.
  • Nothing makes an IT department more nervous than asking for a feed to a key operational system. Moreover, a lot of BI tools are resource hungry. Your requirements should dictate what, how much, and how often (that is, how “real time” you need it to be) data must be fed into your data warehousing technology.
  • In other words, you need one big feed to serve all instead of hundreds of operational, system-killing little feeds that can’t be controlled easily.
  • You'll probably need more than one tool to suit all of your use cases.
  • You did your homework, identified the use cases, picked a good team, started a data integration project, and chose the right tools.
  • Now comes the hard part: changing your business and your decisions based on the data and the reports. Managers, like other human beings, resist change.
  • oreover, BI projects shouldn't have a fixed beginning and end -- this isn't a sprint to become “data driven.”
  • A process is needed
  • and find new opportunities in the data.
  • Here's the bottom line, in a handy do's-and-don'ts format: Don’t simply run a tool-choice project Do cherry-pick the right team Do integrate the data so that it can be queried performance-wise without bringing down the house Don’t merely pick a tool -- pick the right tools for all your requirements and use cases Do let the data change your decision making and the structure of your organization itself if necessary Do have a process to weed out useless analytics and find new ones
cezarovidiu

BI Brief - Four Legs of a Successful Business Intelligence (BI) Project Team - 0 views

  • 1. Project Sponsorship and Governance 2. Project Management 3. Development Team (Core Team) 4. Extended Project Team
  • 1. Project Sponsorship and Governance IT and the business should form a BI steering committee to sponsor and govern design, development, deployment, and ongoing support. It needs both the CIO and a business executive, such as CFO, COO, or a senior VP of marketing/sales to commit budget, time, and resources. The business sponsor needs the project to succeed. The CIO is committed to what is being built and how.
  • 2. Project Management Project management includes managing daily tasks, reporting status, and communicating to the extended project team, steering committee, and affected business users. The project management team needs extensive business knowledge, BI expertise, DW architecture background, and people management, project management, and communications skills. The project management team includes three functions or members: Project development manager - Responsible for deliverables, managing team resources, monitoring tasks, reporting status, and communications. Requires a hands-on IT manager with a background in iterative development. Must understand the changes caused by this approach and the impact on the business, project resources, schedule and the trade-offs. Business advisor - Works within the sponsoring business organization. Responsible for the deliverables of the business resources on the project's extended team. Serves as the business advocate on the project team and the project advocate within the business community. Often, the business advocate is a project co-manager who defers to the IT project manager the daily IT tasks but oversees the budget and business deliverables. BI/DW project advisor - Has enough expertise with architectures and technologies to guides the project team on their use. Ensures that architecture, data models, databases, ETL code, and BI tools are all being used effectively and conform to best practices and standards.
  • ...2 more annotations...
  • 3. Development Team (Core Team) The core project team is divided into four sub-teams: Business requirements - This sub-team may have business people who understand IT systems, or IT people who understand the business. In either case, the team represents the business and their interests. They are responsible for gathering and prioritizing business needs; translating them into IT systems requirements; interacting with the business on the data quality and completeness; and ensuring the business provides feedback on how well the solutions generated meet their needs. BI architecture - Develops the overall BI architecture, selects the appropriate technology, creates the data models, maps the overall data workflow from source systems to BI analytics, and oversees the ETL and BI development teams from a technical perspective. ETL development - Receives the business and data requirements, as well as the target data models to be used by BI analytics. Develops the ETL code needed to gather data from the appropriate source systems into the BI databases. Often, a system analyst who is a expert in the source systems such as SAP is part of the team to provide knowledge of the data sources, customizations, and data quality. BI development - Create the reports or analytics that the business users will interact with to do their jobs. This is often a very iterative process and requires much interaction with the business users.
  • 4. Extended Project Team There are several functions required by the project team that are often accomplished through an "extended" team: Players - A group of business users are signed up to "play with" or test the BI analytics and reports as they are developed to provide feedback to the core development team. This is a virtual team that gets together at specific periods of the project but they are committed to this role during those periods. Testers - A group of resources are gathered, similarly to the virtual team above, to perform more extensive QA testing of the BI analytics, ETL processes, and overall systems testing. You may have project members test other members' work, such as the ETL team test the BI analytics and visa versa. Operators - IT operations is often separated from the development team but it is critical that they are involved from the beginning of the project to ensure that the systems are developed and deployed within your company's infrastructure. Key functions are database administration, systems administration, and networks. In addition, this extended team may also include help desk and training resources if they are usually provided outside of development.
cezarovidiu

Tableau Software's Pat Hanrahan on "What Is a Data Scientist?" - Forbes - 0 views

  • In the contemporary enterprise, almost everyone will need to have data-science skills of some kind.
  • “When most people think of a data scientist, they think of a statistician, a guy with ‘analyst’ in his title,’” Hanrahan says. “Or, someone who works in IT and manages the data warehouses. To do these jobs, you certainly needed programming skills; you probably needed advanced statistics skills, or some combination of those skills.”
  • “At the most basic level, you are a data scientist if you have the analytical skills and the tools to ‘get’ data, manipulate it and make decisions with it,” he says.
  •  
    "What is a Data Scientist?"
cezarovidiu

Universities Offer Courses in a Hot New Field - Data Science - NYTimes.com - 0 views

  • Data scientists are the magicians of the Big Data era. They crunch the data, use mathematical models to analyze it and create narratives or visualizations to explain it, then suggest how to use the information to make decisions.
  • Rachel Schutt, a senior research scientist at Johnson Research Labs, taught “Introduction to Data Science” last semester at Columbia (its first course with “data science” in the title). She described the data scientist this way: “a hybrid computer scientist software engineer statistician.” And added: “The best tend to be really curious people, thinkers who ask good questions and are O.K. dealing with unstructured situations and trying to find structure in them.”
cezarovidiu

Why Soft Skills Matter in Data Science - 0 views

  • You cannot accept problems as handed to you in the business environment. Never allow yourself to be the analyst to whom problems are “thrown over the fence.” Engage with the people whose challenges you’re tackling to make sure you’re solving the right problem. Learn the business’s processes and the data that’s generated and saved. Learn how folks are handling the problem now, and what metrics they use (or ignore) to gauge success.
  • Solve the correct, yet often misrepresented, problem. This is something no mathematical model will ever say to you. No mathematical model can ever say, “Hey, good job formulating this optimization model, but I think you should take a step back and change your business a little instead.” And that leads me to my next point: Learn how to communicate.
  • In today’s business environment, it is often unacceptable to be skilled at only one thing. Data scientists are expected to be polyglots who understand math, code, and the plain-speak (or sports analogy-ridden speak . . . ugh) of business. And the only way to get good at speaking to other folks, just like the only way to get good at math, is through practice.
  • ...4 more annotations...
  • Beware the Three-Headed Geek-Monster: Tools, Performance, and Mathematical Perfection Many things can sabotage the use of analytics within the workplace. Politics and infighting perhaps; a bad experience from a previous “enterprise, business intelligence, cloud dashboard” project; or peers who don’t want their “dark art” optimized or automated for fear that their jobs will become redundant.
  • Not all hurdles are within your control as an analytics professional. But some are. There are three primary ways I see analytics folks sabotage their own work: overly complex modeling, tool obsession, and fixation on performance.
  • In other words, work with the rest of your organization to do better business, not to do data science for its own sake.
  • Data Smart: Using Data Science to Transform Information into Insight by John W. Foreman. Copyright © 2013.
cezarovidiu

What's in a Tag? | ClickZ - 0 views

  • The tag-management industry is growing rapidly, as tags are critical to gathering data about your customers.
  • It's the early days for tag management, but the industry is growing rapidly because it's not so much about tags, but about the bigger challenge of using digital data.
  • Where does tag management fit in the data picture? Here's an example someone shared with me recently: He had gone to an antivirus product's website, read the reviews, and bought the software. In the days that followed, however, he suddenly began to see banner ads from that same software maker whenever he visited CNN, ESPN, and other favorite websites. The software maker knew he had visited its website, but not that he already bought the product. They were retargeting him with banner ads at unnecessary cost and no purpose.
  • ...11 more annotations...
  • Tag management fixes this problem.
  • Most marketing teams struggle with the volume, velocity, and variety of digital data generated every time someone touches the brand. You need insights from the data. You need to understand cross-channel behavior and run predictive "what if" scenarios to improve the effectiveness of your media mix. Tag management can create a foundation to make it easier to use multichannel marketing analytics for these purposes.
  • But one of the big improvements introduced by tag management systems is this: non-technical marketers can do their own tag management.
  • No need to ask IT to deploy tags.
  • You can deploy just one tag, sometimes even just a single line of code, and then manage all the tags through a single user interface.
  • That's a big change from being forced to modify source code on your website.
  • The best tag management systems unite tagged data in one place - automatically.
  • Now the best tag management systems track a data record each time a consumer touches your brand - and deliver it to you in one place.
  • what each consumer has viewed, on what platform, how long they spent with your content, and whether they purchased anything. You get a unified view for everything the consumer has done across all marketing channels.
  • they include the right to be forgotten, easier access to your own data, explicit consent over the use of your data, and privacy by design by default.
  • And, it's clear that the best tag management systems can be a foundation for building those elusive, one-to-one relationships with customers, while using marketing analytics to further improve your marketing decisions about how, when, and where to relate to them.
cezarovidiu

Should your company hire a Chief Data Officer? | Enterprise CIO Forum - 0 views

  • Today, every business is a data business. If you’re a manufacturer of consumer goods, supply chain is absolutely central to what you do, and that’s software. It is all about data. That’s the back end. On the front end, social is absolutely critical if you’re in a consumer-facing business. This makes me recollect Geoffrey Moore’s notion about the implications of systems of engagement for corporations. Systems of engagement—such as Facebook, Google, and so on—generate vast amounts of information about consumer behavior. 
  • And this is why enterprises are hiring CDOs. The information in systems of engagement becomes absolutely critical to large-scale successful consumer businesses. If you’re not leveraging it, then your competitor will outpace you in terms of customer knowledge.
  • CDO wears two hats. On one hand, the CDO is responsible for securing the customer data that’s inside the enterprise. In markets where the privacy laws are extremely strict (such as Germany or Canada), that’s a serious responsibility. At a global company, the CDO must manage consumer data at different levels in different countries in different ways, or creating an enterprise standard for data management globally that reflects the toughest regulation anywhere, which is what HP does.
  • ...1 more annotation...
  • The CDO must also deal with this new opportunity to exploit the much larger range of customer information that exists outside the enterprise. The challenge is to link customer information outside the enterprise with the information that’s inside, and do it in such a way that your data doesn’t leak. You want to run your analytics externally—on Facebook’s platform, for example, because it’s impossible to bring all of Facebook’s customer behavior information into your internal systems. So you never want to pass anything out, but you don’t want to bring everything in. It becomes a game of sophisticated integration.
  •  
    "customer knowledge"
cezarovidiu

Does Excel Power Pivot Replace the Data Warehouse? | SQL Server BI Blog - 0 views

  • Excel Power Pivot is targeted for Personal and Team Business Intelligence (BI) solution use cases.
  • Power Pivot also is excellent for quick prototypes and proofs-of-concept.
  • no row level security,
  • ...2 more annotations...
  • The more advanced features include partitioning for large-scale data sources and role based security.
  • A data mart or data warehouse is often the blessed, single version of the truth since it uses governed, controlled data loading and ETL processes to combine disparate data sources, applies extensive business logic and proven data modeling design patterns that can securely, accurately and efficiently report data changes over time periods.
cezarovidiu

Top Mistakes to Avoid in Analytics Implementations | StatSlice Business Intelligence an... - 0 views

  • Mistake 1.  Not putting a strong interdisciplinary team together. It is impossible to put together an analytics platform without understanding the needs of the customers who will use it.  Sounds simple, right?  Who wouldn’t do that?  You’d be surprised how many analytics projects are wrapped up by IT because “they think” they know the customer needs.  Not assembling the right team is clearly the biggest mistake companies make.  Many times what is on your mind (and if you’re an IT person willing to admit it) is that you are considering converting all those favorite company reports.  Your goal should not be that.  Your goal is to create a system—human engineered with customers, financial people, IT folks, analysts, and others—that give people new and exciting ways to look at information.  It should give you new insights. New competitive information.  If you don’t get the right team put together, you’ll find someone longing for the good old days and their old dusty reports.  Or worse yet, still finding ways to generate those old dusty reports. Mistake 2.  Not having the right talent to design, build, run and update your analytics system.  It is undeniable that there is now high demand for business analytics specialists.  There are not a lot of them out there that really know what to do unless they’ve been burned a few times and have survived and then built successful BA systems.  This is reflected by the fact you see so many analytics vendors offer, or often recommend, third-party consulting and training to help the organization develop their business analytic skills.  Work hard to build a three-way partnership between the vendor, your own team, and an implementation partner.  If you develop those relationships, risk of failure goes way down.
  • Mistake 3.  Putting the wrong kind of analyst or designer on the project. This is somewhat related to Mistake 2 but with some subtle differences.  People have different skillsets so you need to make sure the person you’re considering to put on the project is the right “kind.”  For example, when you put the design together you need both drill-down and summary models.  Both have different types of users.  Does this person know how to do both?  Or, for example, inexperience in an analyst might lead to them believing vendor claims and not be able to verify them as to functionality or time to implement. Mistake 4.  Not understanding how clean the data is you are getting and the time frame to get it clean.  Profile your data to understand the quality of your source data.  This will allow you to adjust your system accordingly to compensate for some of those issues or more importantly push data fixes to your source systems.  Ensure high quality data or your risk upsetting your customers.  If you don’t have a good understanding of the quality of your data, you could easily find yourself way behind schedule even though the actual analytics and business intelligence framework you are building is coming along fine. Mistake 5.  Picking the wrong tools.  How often do organizations buy software tools that just sit on the shelve?  This often comes from management rushing into a quick decision based on a few demos they have seen.  Picking the right analytics tools requires an in-depth understanding of your requirements as well as the strengths and weaknesses of the tools you are evaluating.  The best way to achieve this understanding is by getting an unbiased implementation partner to build a proof of concept with a subset of your own data and prove out the functionality of the tools you are considering. Bottom Line.  Think things through carefully. Make sure you put the right team together.  Have a data cleansing plan.  If the hype sounds too good to be true—have someone prove it to you.
cezarovidiu

Business Intelligence Blog - The ElastiCube Chronicles - 0 views

  • SiSense’s survey finds that salaries for data professionals are on the rise across all geographies. The annual earnings of a data professional can range from an average of $55,000 USD for a data analyst to an average of $132,000 for VP Analytics. As many as 61% of the survey respondents reported higher earnings in 2012 compared to 2011, and only 12% reported lower earnings.
  • Other highlights of the survey findings include: Data professionals are highly educated. 85% of the respondents have some college degree, 39% have a Master’s degree, and 5% are Ph.D.’s. Those with doctoral degrees earn on average 65% more than those with Master’s degrees, who in turn earn 16% more than those with Bachelor’s degrees. On the job experience is even more important than education in determining salary levels. On average, professionals with ten or more years of experience earn 80% more than those with 3 years or less.
  • At the same time, the survey shows that those with 6 years or less make up as much as 59% of the data profession workforce.
  • ...1 more annotation...
  • Most Data Professionals work in teams of up to five people. “Companies are starting to realize that Data is key to their success. The majority of them, though are not growing their Data Science teams fast enough to win. This maybe because they don’t want to or because they can’t. This is an alarming trend though and only software can come to the rescue,” noted Aziza.
cezarovidiu

Magic Quadrant for Data Warehouse Database Management Systems - 0 views

  • relational database management systems (DBMSs) used as platforms for data warehouses
  • It is important to note that a DBMS does not in itself constitute a data warehouse — rather, a data warehouse can be deployed on a DBMS platform.
  • a data warehouse is simply a warehouse of data, not a specific class or type of technology
  •  
    "Magic Quadrant for Data Warehouse Database Management Systems"
cezarovidiu

Don't have a big data strategy yet? Good! | ZDNet - 0 views

  • Business Intelligence And Big Data, Q4 2012 survey of 634 BI users and planners, we found that social and mobile data is actually a pretty low priority: While big data technologies like Hadoop can definitely deal with over-hyped new data like mobile and social, the broad demand is simply not there yet. Instead we find leaders recognize that the big deal about big data is the potential for getting more value more quickly from more data, at a lower cost and with greater agility; and it is a whole range of technologies and new techniques that make more possible.
1 - 20 of 142 Next › Last »
Showing 20 items per page