Skip to main content

Home/ BI-TAGS/ Group items tagged data model

Rss Feed Group items tagged

cezarovidiu

What is business intelligence (BI)? - Definition from WhatIs.com - 0 views

  • Business intelligence is a data analysis process aimed at boosting business performance by helping corporate executives and other end users make more informed decisions.
  • Business intelligence (BI) is a technology-driven process for analyzing data and presenting actionable information to help corporate executives, business managers and other end users make more informed business decisions.
  • BI encompasses a variety of tools, applications and methodologies that enable organizations to collect data from internal systems and external sources, prepare it for analysis, develop and run queries against the data, and create reports, dashboards and data visualizations to make the analytical results available to corporate decision makers as well as operational workers.
  • ...9 more annotations...
  • The potential benefits of business intelligence programs include accelerating and improving decision making; optimizing internal business processes; increasing operational efficiency; driving new revenues; and gaining competitive advantages over business rivals. BI systems can also help companies identify market trends and spot business problems that need to be addressed.
  • BI data can include historical information, as well as new data gathered from source systems as it is generated, enabling BI analysis to support both strategic and tactical decision-making processes.
  • BI programs can also incorporate forms of advanced analytics, such as data mining, predictive analytics, text mining, statistical analysis and big data analytics.
  • In many cases though, advanced analytics projects are conducted and managed by separate teams of data scientists, statisticians, predictive modelers and other skilled analytics professionals, while BI teams oversee more straightforward querying and analysis of business data.
  • Business intelligence data typically is stored in a data warehouse or smaller data marts that hold subsets of a company's information. In addition, Hadoop systems are increasingly being used within BI architectures as repositories or landing pads for BI and analytics data, especially for unstructured data, log files, sensor data and other types of big data. Before it's used in BI applications, raw data from different source systems must be integrated, consolidated and cleansed using data integration and data quality tools to ensure that users are analyzing accurate and consistent information.
  • In addition to BI managers, business intelligence teams generally include a mix of BI architects, BI developers, business analysts and data management professionals; business users often are also included to represent the business side and make sure its needs are met in the BI development process.
  • To help with that, a growing number of organizations are replacing traditional waterfall development with Agile BI and data warehousing approaches that use Agile software development techniques to break up BI projects into small chunks and deliver new functionality to end users on an incremental and iterative basis.
  • consultant Howard Dresner is credited with first proposing it in 1989 as an umbrella category for applying data analysis techniques to support business decision-making processes.
  • Business intelligence is sometimes used interchangeably with business analytics; in other cases, business analytics is used either more narrowly to refer to advanced data analytics or more broadly to include both BI and advanced analytics.
cezarovidiu

Rittman Mead Consulting - The Changing World of Business Intelligence - 0 views

  • Schema on write This is the traditional approach for Business Intelligence. A model, often dimensional, is built as part of the design process. This model is an abstraction of the complexity of the underlying systems, put in business terms. The purpose of the model is to allow the business users to interrogate the data in a way they understand.
  • The model is instantiated through physical database tables and the date is loaded through an ETL (extract, transform and load) process that takes data from one or more source systems and transforms it to fit the model, then loads it into the model.
  • The key thing is that the model is determined before the data is finally written and the users are very much guided or driven by the model in how they query the data and what results they can get from the system. The designer must anticipate the queries and requests in advance of the user asking the questions.
  • ...3 more annotations...
  • Schema on read Schema on read works on a different principle and is more common in the Big Data world. The data is not transformed in any way when it is stored, the data store acts as a big bucket. The modelling of the data only occurs when the data is read. Map/Reduce is the clearest example, the mapping is the understanding of the data structure. Hadoop is a large distributed file system, which is very good at storing large volumes of data, this is potential. It is only the mapping of this data that provides value, this is done when the data is read, not written.
  • New World Order So whereas Business Intelligence used to always be driven by the model, the ETL process to populate the model and the reporting tool to query the model, there is now an approach where the data is collected its raw form, and advanced statistical or analytical tools are used to interrogate the data. An example of one such tool is R.
  • The driver for which approach to use is often driven by what the user wants to find out. If the question is clearly formed and the sources of data that are required to answer it well understood, for example how many units of a product have we sold, then the traditional schema on write approach is best.
cezarovidiu

Focus on Valuable Data - Not Big Data - to Boost Conversions and ROI | ClickZ - 0 views

  • Big Data has been all the rage. But fast data, even if it is small, can be more valuable than complicated masses of information.
  • Here's why: All the focus on "bigger is better" has overlooked the fact that most Big Data segments have not been validated with a business application or value.
  • Those kinds of analytics can help you find the right streams to access and work with, and also can help you build out robust programs that identify valuable customers.
  • ...1 more annotation...
  • 1) Your First-Party Data: The primary and most valuable data set you can access, first-party data encompasses transactional and other customer-level profile information you have on your customers. It could also include your own off-line segmentation analysis that allows you to map a customer to a customer profile around which you build your marketing programs. This can also include your analytics or other on-site tracking data, which can deliver behavioral insight to your consumers. This data can be difficult to export from its current environment due to the ad hoc nature of the data, but, if possible, look at ways to make this information accessible to your digital sites. 2) Third-Party Data: A consumer's broader Web browsing and buying history can now be accessed in session to provide you with more context on their likes and habits. Data management platforms (DMPs) and other data aggregators are accelerating this offering and, just as importantly, the availability of this type of data. This is invaluable in the context of new visitors who you know nothing about historically. 3) Real-Time Behavior: Let's not forget what our customers are telling us with each click. We get enamored with our predictive modeling to the point that we do not see the tell-tale signs as they are happening. Take the time to stop, look, and react. Your analytic tools, personalization tools, and other software-as-a-service (SaaS) platforms can help you trigger alternate site experiences based on every click you see.
cezarovidiu

Magic Quadrant for Business Intelligence and Analytics Platforms - 0 views

  • Integration BI infrastructure: All tools in the platform use the same security, metadata, administration, portal integration, object model and query engine, and should share the same look and feel. Metadata management: Tools should leverage the same metadata, and the tools should provide a robust way to search, capture, store, reuse and publish metadata objects, such as dimensions, hierarchies, measures, performance metrics and report layout objects. Development tools: The platform should provide a set of programmatic and visual tools, coupled with a software developer's kit for creating analytic applications, integrating them into a business process, and/or embedding them in another application. Collaboration: Enables users to share and discuss information and analytic content, and/or to manage hierarchies and metrics via discussion threads, chat and annotations.
  • Information Delivery Reporting: Provides the ability to create formatted and interactive reports, with or without parameters, with highly scalable distribution and scheduling capabilities. Dashboards: Includes the ability to publish Web-based or mobile reports with intuitive interactive displays that indicate the state of a performance metric compared with a goal or target value. Increasingly, dashboards are used to disseminate real-time data from operational applications, or in conjunction with a complex-event processing engine. Ad hoc query: Enables users to ask their own questions of the data, without relying on IT to create a report. In particular, the tools must have a robust semantic layer to enable users to navigate available data sources. Microsoft Office integration: Sometimes, Microsoft Office (particularly Excel) acts as the reporting or analytics client. In these cases, it is vital that the tool provides integration with Microsoft Office, including support for document and presentation formats, formulas, data "refreshes" and pivot tables. Advanced integration includes cell locking and write-back. Search-based BI: Applies a search index to structured and unstructured data sources and maps them into a classification structure of dimensions and measures that users can easily navigate and explore using a search interface. Mobile BI: Enables organizations to deliver analytic content to mobile devices in a publishing and/or interactive mode, and takes advantage of the mobile client's location awareness.
  • Analysis Online analytical processing (OLAP): Enables users to analyze data with fast query and calculation performance, enabling a style of analysis known as "slicing and dicing." Users are able to navigate multidimensional drill paths. They also have the ability to write back values to a proprietary database for planning and "what if" modeling purposes. This capability could span a variety of data architectures (such as relational or multidimensional) and storage architectures (such as disk-based or in-memory). Interactive visualization: Gives users the ability to display numerous aspects of the data more efficiently by using interactive pictures and charts, instead of rows and columns. Predictive modeling and data mining: Enables organizations to classify categorical variables, and to estimate continuous variables using mathematical algorithms. Scorecards: These take the metrics displayed in a dashboard a step further by applying them to a strategy map that aligns key performance indicators (KPIs) with a strategic objective. Prescriptive modeling, simulation and optimization: Supports decision making by enabling organizations to select the correct value of a variable based on a set of constraints for deterministic processes, and by modeling outcomes for stochastic processes.
  • ...7 more annotations...
  • These capabilities enable organizations to build precise systems of classification and measurement to support decision making and improve performance. BI and analytic platforms enable companies to measure and improve the metrics that matter most to their businesses, such as sales, profits, costs, quality defects, safety incidents, customer satisfaction, on-time delivery and so on. BI and analytic platforms also enable organizations to classify the dimensions of their businesses — such as their customers, products and employees — with more granular precision. With these capabilities, marketers can better understand which customers are most likely to churn. HR managers can better understand which attributes to look for when recruiting top performers. Supply chain managers can better understand which inventory allocation levels will keep costs low without increasing out-of-stock incidents.
  • descriptive, diagnostic, predictive and prescriptive analytics
  • "descriptive"
  • diagnostic
  • data discovery vendors — such as QlikTech, Salient Management Company, Tableau Software and Tibco Spotfire — received more positive feedback than vendors offering OLAP cube and semantic-layer-based architectures.
  • Microsoft Excel users are often disaffected business BI users who are unable to conduct the analysis they want using enterprise, IT-centric tools. Since these users are the typical target users of data discovery tool vendors, Microsoft's aggressive plans to enhance Excel will likely pose an additional competitive threat beyond the mainstreaming and integration of data discovery features as part of the other leading, IT-centric enterprise platforms.
  • Building on the in-memory capabilities of PowerPivot in SQL Server 2012, Microsoft introduced a fully in-memory version of Microsoft Analysis Services cubes, based on the same data structure as PowerPivot, to address the needs of organizations that are turning to newer in-memory OLAP architectures over traditional, multidimensional OLAP architectures to support dynamic and interactive analysis of large datasets. Above-average performance ratings suggest that customers are happy with the in-memory improvements in SQL Server 2012 compared with SQL Server 2008 R2, which ranks below the survey average.
  •  
    "Gartner defines the business intelligence (BI) and analytics platform market as a software platform that delivers 15 capabilities across three categories: integration, information delivery and analysis."
cezarovidiu

Why Soft Skills Matter in Data Science - 0 views

  • You cannot accept problems as handed to you in the business environment. Never allow yourself to be the analyst to whom problems are “thrown over the fence.” Engage with the people whose challenges you’re tackling to make sure you’re solving the right problem. Learn the business’s processes and the data that’s generated and saved. Learn how folks are handling the problem now, and what metrics they use (or ignore) to gauge success.
  • Solve the correct, yet often misrepresented, problem. This is something no mathematical model will ever say to you. No mathematical model can ever say, “Hey, good job formulating this optimization model, but I think you should take a step back and change your business a little instead.” And that leads me to my next point: Learn how to communicate.
  • In today’s business environment, it is often unacceptable to be skilled at only one thing. Data scientists are expected to be polyglots who understand math, code, and the plain-speak (or sports analogy-ridden speak . . . ugh) of business. And the only way to get good at speaking to other folks, just like the only way to get good at math, is through practice.
  • ...4 more annotations...
  • Beware the Three-Headed Geek-Monster: Tools, Performance, and Mathematical Perfection Many things can sabotage the use of analytics within the workplace. Politics and infighting perhaps; a bad experience from a previous “enterprise, business intelligence, cloud dashboard” project; or peers who don’t want their “dark art” optimized or automated for fear that their jobs will become redundant.
  • Not all hurdles are within your control as an analytics professional. But some are. There are three primary ways I see analytics folks sabotage their own work: overly complex modeling, tool obsession, and fixation on performance.
  • In other words, work with the rest of your organization to do better business, not to do data science for its own sake.
  • Data Smart: Using Data Science to Transform Information into Insight by John W. Foreman. Copyright © 2013.
cezarovidiu

BI Brief - Four Legs of a Successful Business Intelligence (BI) Project Team - 0 views

  • 1. Project Sponsorship and Governance 2. Project Management 3. Development Team (Core Team) 4. Extended Project Team
  • 1. Project Sponsorship and Governance IT and the business should form a BI steering committee to sponsor and govern design, development, deployment, and ongoing support. It needs both the CIO and a business executive, such as CFO, COO, or a senior VP of marketing/sales to commit budget, time, and resources. The business sponsor needs the project to succeed. The CIO is committed to what is being built and how.
  • 2. Project Management Project management includes managing daily tasks, reporting status, and communicating to the extended project team, steering committee, and affected business users. The project management team needs extensive business knowledge, BI expertise, DW architecture background, and people management, project management, and communications skills. The project management team includes three functions or members: Project development manager - Responsible for deliverables, managing team resources, monitoring tasks, reporting status, and communications. Requires a hands-on IT manager with a background in iterative development. Must understand the changes caused by this approach and the impact on the business, project resources, schedule and the trade-offs. Business advisor - Works within the sponsoring business organization. Responsible for the deliverables of the business resources on the project's extended team. Serves as the business advocate on the project team and the project advocate within the business community. Often, the business advocate is a project co-manager who defers to the IT project manager the daily IT tasks but oversees the budget and business deliverables. BI/DW project advisor - Has enough expertise with architectures and technologies to guides the project team on their use. Ensures that architecture, data models, databases, ETL code, and BI tools are all being used effectively and conform to best practices and standards.
  • ...2 more annotations...
  • 3. Development Team (Core Team) The core project team is divided into four sub-teams: Business requirements - This sub-team may have business people who understand IT systems, or IT people who understand the business. In either case, the team represents the business and their interests. They are responsible for gathering and prioritizing business needs; translating them into IT systems requirements; interacting with the business on the data quality and completeness; and ensuring the business provides feedback on how well the solutions generated meet their needs. BI architecture - Develops the overall BI architecture, selects the appropriate technology, creates the data models, maps the overall data workflow from source systems to BI analytics, and oversees the ETL and BI development teams from a technical perspective. ETL development - Receives the business and data requirements, as well as the target data models to be used by BI analytics. Develops the ETL code needed to gather data from the appropriate source systems into the BI databases. Often, a system analyst who is a expert in the source systems such as SAP is part of the team to provide knowledge of the data sources, customizations, and data quality. BI development - Create the reports or analytics that the business users will interact with to do their jobs. This is often a very iterative process and requires much interaction with the business users.
  • 4. Extended Project Team There are several functions required by the project team that are often accomplished through an "extended" team: Players - A group of business users are signed up to "play with" or test the BI analytics and reports as they are developed to provide feedback to the core development team. This is a virtual team that gets together at specific periods of the project but they are committed to this role during those periods. Testers - A group of resources are gathered, similarly to the virtual team above, to perform more extensive QA testing of the BI analytics, ETL processes, and overall systems testing. You may have project members test other members' work, such as the ETL team test the BI analytics and visa versa. Operators - IT operations is often separated from the development team but it is critical that they are involved from the beginning of the project to ensure that the systems are developed and deployed within your company's infrastructure. Key functions are database administration, systems administration, and networks. In addition, this extended team may also include help desk and training resources if they are usually provided outside of development.
cezarovidiu

Universities Offer Courses in a Hot New Field - Data Science - NYTimes.com - 0 views

  • Data scientists are the magicians of the Big Data era. They crunch the data, use mathematical models to analyze it and create narratives or visualizations to explain it, then suggest how to use the information to make decisions.
  • Rachel Schutt, a senior research scientist at Johnson Research Labs, taught “Introduction to Data Science” last semester at Columbia (its first course with “data science” in the title). She described the data scientist this way: “a hybrid computer scientist software engineer statistician.” And added: “The best tend to be really curious people, thinkers who ask good questions and are O.K. dealing with unstructured situations and trying to find structure in them.”
cezarovidiu

Does Excel Power Pivot Replace the Data Warehouse? | SQL Server BI Blog - 0 views

  • Excel Power Pivot is targeted for Personal and Team Business Intelligence (BI) solution use cases.
  • Power Pivot also is excellent for quick prototypes and proofs-of-concept.
  • no row level security,
  • ...2 more annotations...
  • The more advanced features include partitioning for large-scale data sources and role based security.
  • A data mart or data warehouse is often the blessed, single version of the truth since it uses governed, controlled data loading and ETL processes to combine disparate data sources, applies extensive business logic and proven data modeling design patterns that can securely, accurately and efficiently report data changes over time periods.
cezarovidiu

Top Mistakes to Avoid in Analytics Implementations | StatSlice Business Intelligence an... - 0 views

  • Mistake 1.  Not putting a strong interdisciplinary team together. It is impossible to put together an analytics platform without understanding the needs of the customers who will use it.  Sounds simple, right?  Who wouldn’t do that?  You’d be surprised how many analytics projects are wrapped up by IT because “they think” they know the customer needs.  Not assembling the right team is clearly the biggest mistake companies make.  Many times what is on your mind (and if you’re an IT person willing to admit it) is that you are considering converting all those favorite company reports.  Your goal should not be that.  Your goal is to create a system—human engineered with customers, financial people, IT folks, analysts, and others—that give people new and exciting ways to look at information.  It should give you new insights. New competitive information.  If you don’t get the right team put together, you’ll find someone longing for the good old days and their old dusty reports.  Or worse yet, still finding ways to generate those old dusty reports. Mistake 2.  Not having the right talent to design, build, run and update your analytics system.  It is undeniable that there is now high demand for business analytics specialists.  There are not a lot of them out there that really know what to do unless they’ve been burned a few times and have survived and then built successful BA systems.  This is reflected by the fact you see so many analytics vendors offer, or often recommend, third-party consulting and training to help the organization develop their business analytic skills.  Work hard to build a three-way partnership between the vendor, your own team, and an implementation partner.  If you develop those relationships, risk of failure goes way down.
  • Mistake 3.  Putting the wrong kind of analyst or designer on the project. This is somewhat related to Mistake 2 but with some subtle differences.  People have different skillsets so you need to make sure the person you’re considering to put on the project is the right “kind.”  For example, when you put the design together you need both drill-down and summary models.  Both have different types of users.  Does this person know how to do both?  Or, for example, inexperience in an analyst might lead to them believing vendor claims and not be able to verify them as to functionality or time to implement. Mistake 4.  Not understanding how clean the data is you are getting and the time frame to get it clean.  Profile your data to understand the quality of your source data.  This will allow you to adjust your system accordingly to compensate for some of those issues or more importantly push data fixes to your source systems.  Ensure high quality data or your risk upsetting your customers.  If you don’t have a good understanding of the quality of your data, you could easily find yourself way behind schedule even though the actual analytics and business intelligence framework you are building is coming along fine. Mistake 5.  Picking the wrong tools.  How often do organizations buy software tools that just sit on the shelve?  This often comes from management rushing into a quick decision based on a few demos they have seen.  Picking the right analytics tools requires an in-depth understanding of your requirements as well as the strengths and weaknesses of the tools you are evaluating.  The best way to achieve this understanding is by getting an unbiased implementation partner to build a proof of concept with a subset of your own data and prove out the functionality of the tools you are considering. Bottom Line.  Think things through carefully. Make sure you put the right team together.  Have a data cleansing plan.  If the hype sounds too good to be true—have someone prove it to you.
cezarovidiu

Is Big Data Really Working for Marketers? | ClickZ - 0 views

  • Channel Optimization. Many marketers struggle to optimize each individual channel, let alone optimizing at a customer level across many channels. To the extent that Big Data can help marketers understand what is important in the moment and across touch points, that could be valuable, but it seems more of us need stronger attribution models and analytics methodologies more than access to data. Big Data does seem to be valuable if you want to understand which customers are highest value within each channel and across channels, because the platforms that manage Big Data can handle both structured and unstructured data - which is what you need to truly include Web/clickstream and social data in your analysis.
cezarovidiu

Rittman Mead Consulting » Blog Archive » Event Triggers in BI Publisher 11g - 0 views

  •  
    "Event Triggers in BI Publisher 11g December 20th, 2011 by Robin Moffatt Event Triggers in BI Publisher 11g give the facility to call a function in Oracle either before or after a data set is refreshed. The function must return a boolean (true/false), and if it returns false the data model will abort execution."
cezarovidiu

Rittman Mead Consulting » Blog Archive » Using OBIEE against Transactional Sc... - 0 views

  • The best practice in business intelligence delivery is always to build a data warehouse.
  • Pure transactional reporting is problematic. There are, of course, the usual performance issues. Equally troublesome is the difficulty in distilling a physical model down to a format that is easy for business users to understand. Dimensional models are typically the way business users envision their business: simple, inclusive structures for each entity. The standard OLTP data model that takes two of the four walls in the conference room to display will never make sense to your average business user.
cezarovidiu

Google Reader (250) - 0 views

  • What this means in practice is that when the BI Server component starts up, it creates and reserves a number of threads in advance, determined by a number of parameters including SERVER_THREAD_RANGE.
  • You can see these threads running and ready to perform tasks for the BI Server component by using a tool such as Process Explorer for Windows
  • Thinking it through a bit, any given single query is, to a certain extent, only really going to use a small part of the total amount of CPUs available on a server, because it’s not the BI Server that runs queries in parallel, it’s the underlying database. For example, a single analysis against a single Oracle Database datasource would only really need a single BI Server thread to handle the query request, but when the underlying database receives the query, it might use a large number of its CPUs to process the query, returning results back to the BI Server to then pass back to the Presentation Server for display to the user.
  • ...2 more annotations...
  • The BI Server wouldn’t have any use for any more query threads, as it can’t really do anything with them – the exception to this being queries that generate multiple physical SQLs, for example to join data from multiple sources together and return a single set of data to the user, for which the BI Server could benefit from a higher CPU count if each of these queries in turn led to lots of threads being used – but two queries, in themselves, don’t neccessarily require two CPUs, because of course the BI Server, and the underlying CPUs, are themselves multi-threaded.
  • To conclude then – all things begin equal, the BI Server should make use of all of the CPUs that the underlying operating system presents to it, with the OS itself deciding what threads are scheduled against which CPUs. In-theory, all CPUs on the server are available to each BI Server component, but each OS is different and it might be worth experimenting if you’re sure that certain CPUs aren’t being used – but this is most probably unlikely and the main reason you’d really consider vertical scale-out of BI Server components is for fault-tolerance, or if you’re using a 32-bit OS and each process can only see a subset of the total overall memory. And, bear in mind that however many CPUs the BI Server has available to it, for queries that send just a single SQL statement down to the underlying database server, adding more CPUs or faster CPUs isn’t going to help as only a single (or so) thread will be needed to send the query from the BI Server to the database, and it’s the database that’s doing all of the work – all that this would help with is compilation and post-aggregation work, and enabling the server to handle a higher number of concurrent users. Invest in a better underlying database instead, sort out your data model, and make sure your data source back-end is as optimised as possible.
cezarovidiu

Installing Hadoop for Fedora & Oracle Linux(Single Node Cluster) | accretion infinity - 0 views

  • Hadoop is a framework written in Java for running applications on large clusters of commodity hardware and incorporates features similar to those of the Google File System (GFS) and of the Map Reduce computing paradigm. Hadoop’s HDFS is a highly fault-tolerant distributed file system and, like Hadoop in general, designed to be deployed on low-cost hardware. It provides high throughput access to application data and is suitable for applications that have large data sets.
  • Some of the Hadoop projects we will talk about are: HDFS : A distributed filesystem that runs on large clusters of commodity machines. Map Reduce: A distributed data processing model and execution environment that runs on large clusters of commodity machines. Pig: A data flow language and execution environment for exploring very large datasets. Pig runs on HDFS and MapReduce clusters. HBase: A distributed, column-oriented database. HBase uses HDFS for its underlying storage, and supports both batch-style computations using MapReduce and point queries (random reads). ZooKeeper: A distributed, highly available coordination service. ZooKeeper provides primitives such as distributed locks that can be used for building distributed applications. Oozie: Oozie is a workflow scheduler system to manage Apache Hadoop jobs.
  • Oracle Linux as the operating system and Hadoop 1.1.2 or 1.2.0
cezarovidiu

What Skills Does an Oracle BI Developer Need in 2011? - 0 views

  • OBIEE 11g skills, both in terms of new functionality (mapping, analyses, KPIs and Scorecards etc) and new infrastructure (WebLogic, EM, OPSS etc) A smattering of Essbase skills, focused mainly on the integration with OBIEE and Essbase (and the many workarounds and gotchas) Good ODI skills, both in terms of the basics, but also being able to write knowledge modules, integrate with OBIEE, deployment and migration Solid database skills – OBIEE gave the illusion through aggregates etc that database tuning was redundant, but time has shown it’s by far the biggest success factor in a project – get the database design and optimisation wrong, and your project is toast. You need to know partitioning, materialized views, index types, and increasingly, you need to get yourself on an Exadata project as customers are buying the technology but you can’t teach it to yourself at home BI Apps skills, but watch out for everything changing when BI Apps 11g comes out, and be prepared to learn the Fusion Apps and JDeveloper if you want to stay in the game Looking to the future, keep an eye on technologies such as in-memory (TimesTen), mid-tier caching (Coherence), plus technologies such as Business Activity Monitoring (BAM), “big data” (Hadoop, large data sets, NoSQL), complex event processing and maybe products such as Qlikview, just in case Oracle buys them, or at least to know what the competition are up to, or more importantly pitching to your boss
  • The other thing to bear in mind of course, if you’re an Oracle BI developer, is that you need to have great business, communication and data modeling skills.
cezarovidiu

Successful Social Marketing is So Much More Than Social Media | ClickZ - 0 views

  • In the past, prospects primarily accessed information about a company by interacting directly with a salesperson.
  • As media evolved, mass ads, events, direct mail, and more recently, email, have been the primary tools for engagement.
  • Given the number of consumers posting, blogging, tweeting, liking and sharing, the question for marketers is no longer, Should I use social? It's, How do I use social to its full potential?
  • ...14 more annotations...
  • Social channels are inherently built for sharing and engagement, making them the perfect place to cultivate valuable business relationships. Integrating social into every marketing campaign you run can move you from a company-to-buyer marketing model to a peer-to-peer influence model. This not only builds trust and brand loyalty, but also positively impacts ROI.
  • It can be tempting to jump right in to all the social media sites out there and start posting away. However, before you publish that first nugget of social marketing content, you need to develop your plan.
  • goals and metrics
  • Build a team that is willing and able to dedicate adequate time to social media endeavors.
  • Many marketers fall into the trap of thinking that social media campaigns can be dealt with on an ad hoc basis, but this couldn't be further from the truth. You don't want your company's online personality to come across as erratic or disjointed, so create a policy that guides those who are participating in the social marketing effort and be sure those guidelines are enforced.
  • Once everyone is on board, encourage them to create engaging content. A good starting place is to ask your team members to answer some of the most frequently asked questions they receive on the various social channels. If everyone is a content creator, you'll never be short of ideas.
  • Word-of-mouth is incredibly powerful and the "share" button on every social media channel allows you to tap into millions of different networks. One of the best ways to interact with your audience is by giving them content they genuinely want to share with their networks. Peer recommendation is extremely valuable because people believe their friends much more readily than a company or marketer.
  • A "Refer-a-Friend" campaign promotes a compelling offer via email marketing and social networks, then grants access to special offers for both the referrers and those referred. Using these campaigns will allow you to gather important metrics, like tracking who the biggest influencers are.
  • A "Social Sweepstakes" campaign allows your entrants to spread the word on your behalf. Through the sweepstakes entry, you gain important user data like who is sharing and where they are sharing most.
  • Finally, a "Flash Deal" campaign is similar to Groupon. Flash deals offer a limited amount of deals for a specific time period through your social platforms. If you use these campaigns, be sure to let participants track the deal's progress! These campaigns are fun and viral ways to spread brand awareness and boost new customer numbers with sharing.
  • make sure your shares are measurable. Monitoring social share numbers is not only an easy way to tell what's working and what's not, but also allows you to see your ROI by showing how far your social reach is in relation to how much time and resources you've put in.
  • Google Alerts and search functions, or enterprise level software like Viral Heat or Radian6.
  • Once you hear what people are saying, you can engage them with relevant responses.
  • Social has evolved into much more than just a channel or tactic and should be an ever-present strategy in all aspects of your marketing. Ultimately, if you come up with a plan, encourage creative content, incorporate social marketing into every stage of your funnel, and measure your results, you'll start to see your social efforts move the ROI needle in the right direction.
1 - 18 of 18
Showing 20 items per page