Skip to main content

Home/ Groups/ ODI Data Infrastructure, Institutions & Data Access Initiatives
Ben Snaith

HowToFundTech2020.pdf - 0 views

shared by Ben Snaith on 04 Aug 20 - No Cached
Ben Snaith

Every day, we rely on digital infrastructure built by volunteers. What happens when it ... - 0 views

  • Free and public code grew in direct response to the perceived failings of expensive, proprietary commercial software. As a result, the heart of the problem with digital infrastructure is also part of what makes it so rich with potential: It is not centralized. There is no one person or entity deciding what’s needed and what’s not. There is also no one overseeing how digital infrastructure is implemented. And because the community of volunteers developing this infrastructure has a complicated relationship with what might be seen as a more traditional, or official, way of doing things, few digital infrastructure projects have a clear business model or source of revenue. Even projects that have grown to be used by millions of people tend to lack a cohesive structure and plan for sustaining the technology’s long-term development.
  • We need to start by educating people who are in positions to provide support. Many of them—from start-up engineers to government officials—don’t know enough about how digital infrastructure functions and what it requires, or are under the perception that public software doesn’t need support.
Sonia Duarte

Our services | BSI - 0 views

  •  
    "BSI certification shows the world that you work in the smartest, most efficient ways and that you are continually improving your performance."
Sonia Duarte

How have National Statistical Institutes improved quality in the last 25 years? - IOS P... - 0 views

  • There are still major efforts needed to continuously improve. More focus needs to be put on measuring internal processes, costs, and components of quality other than accuracy. Documentation needs to be regularly updated, methods for incorporating Big Data developed, and flexibility improved so that adaptive methods based on paradata can be used.
  • it takes regular management involvement and procedures to be in place for it to succeed
  • Measurements are vital, but they are not the goal. This will require re-focusing on improving internal processes. It also implies recognizing the need to track costs as a component of quality.
  • ...1 more annotation...
  • While it will continue to be important for NSIs to increase their use of administrative data and the many sources of Big Data, these will rarely be able to be used as stand-alone sources. More often these sources will need to be combined with well-designed survey data to produce a blended, improved product.
  •  
    "Principles and Practices for a Federal Statistical Agency"
Sonia Duarte

How do different communities create unique identifiers? - Lost Boy - 0 views

  • They play an important role, helping to publish, structure and link together data.
  • The simplest way to generate identifiers is by a serial number.
  • the Ordnance Survey TOID identifier is a serial number that looks like this: osgb1000006032892. UPRNs are similar.
  • ...15 more annotations...
  • Some serial numbering systems include built in error-checking to deal with copying errors, using a check digit.
  • The second way of providing unique identifiers is using a name or code.
  • These are typically still assigned by a central authority, sometimes known as a registration agency, but they are constructed in different ways.
  • Encoding information about geography and hierarchy within codes can be useful. It can make them easier to validate.
  • It also mean you can also manipulate them,
  • But encoding lots of information in identifiers also has its downsides. The main one being dealing with changes to administrative areas that mean the hierarchy has changed. Do you reassign all the identifiers?
  • some identifier systems look at reducing the burden on that central authority.
  • federated assignment. This is where the registration agency shares the work of assigning identifiers with other organisations.
  • Another approach to reducing dependence on, and coordination with a single registration agency, is to use what I’ll call “local assignment“.
  • A simplistic approach to local assignment is “block allocation“: handing out blocks of pregenerated identifiers to organisations which can locally assign them.
  • Here the registration agency still generates the identifiers, but the assignment of identifier to “thing” is done locally.
  • A more common approach is to use “prefix allocation“. In this approach the registration agency assigns individual organisations a prefix within the identifier system.
  • One challenge with prefix allocation is ensuring that the rules for locally assigned suffixes work in every context where the identifier needs to appear.
  • In distributed assignment of identifiers, anyone can create an identifier. Rather than requesting an identifier, or a prefix from a registration agency, these systems operate by agreeing rules for how unique identifiers can be constructed.
  • A hash based identifier takes some properties of the thing we want to identify and then use that to construct an identifier. 
Ben Snaith

345725803-The-state-of-weather-data-infrastructure-white-paper.pdf - 1 views

  • From its early beginnings over 150 years ago, weather forecasting at the Met Office has been driven by data. Simple observations recorded and used to hand-plot synoptic charts have been exchanged for the billions of observations received and handled every day, mainly from satellites but also from weather stations, radar , ocean buoys, planes, shipping and the public.
  • The key stages of the weather data value chain are as follows: Ÿ Monitoring and observation of the weather and environment, e.g. by NMSs. Ÿ Numerical weather prediction (NWP) and climate modelling carried out by NMSs to create global, regional and limited area weather forecasts. Private companies are growing their presence in the market and challenging the traditional role of NMSs to provide forecasts to the public, by statistically blending data from NMS models to create their own forecast models, for example. Other companies providing data via online channels and/or apps include The Weather Company, Accuweather or the Climate Corporation. Ÿ Communication and dissemination of forecasts by news, NMS and media organisations like the BBC, Yahoo and Google, or within consumer-targeted mobile and web applications. Ÿ Decision making by individuals and businesses across a variety of sectors, which draws on weather data and reporting.
  • The core data asset of our global weather data infrastructure is observation data that captures a continuous record of weather and climate data around the world. This includes temperature, rainfall, wind speed and details of a host of other atmospheric, surface and marine conditions.
  • ...5 more annotations...
  • The collection of observation data is a global effort. The Global Observing System consists of around 11,000 ground-based monitoring stations supplemented with thousands of sensors installed on weather balloons, aircraft and ships. 3 Observations are also collected from a network of radar installations and satellite-based sensors. As we see later, the ‘official’ observation system is increasingly being supplemented with new sources of observation data from the Internet of Things (IoT).
  • Ensemble model forecasts aim to give an indication of the range of possible future states of the atmosphere and oceans (which are a significant driver of the weather in the atmosphere). This overcomes errors introduced by using imperfect measurement of initial starting conditions that are then amplified by the chaotic nature of the atmosphere. Increasing the number of forecast members over a global scale and at higher resolutions result in data volumes increasing exponentially .
  • Created in 1950, The World Meteorological Organisation (WMO) is made up of 191 member states and territories around the world. The WMO was founded on the principle that global coordination was necessary to reap the benefits of weather and climate data. This includes a commitment to weather data and products being freely exchanged around the world (Obasi, 1995).
  • While the WMO has a global outlook, its work is supplemented by regional meteorological organisations like the European Centre for Medium Range Weather Forecasts (ECMWF) and NMSs, such as the Met Office in the UK
  • There are increasing new sources of weather observation data. In recent years, services like Weather Underground and the Met Office’s Weather Observation Website have demonstrated the potential for people around the world to contribute weather observations about their local areas – using low-cost home weather stations and sensors, for example. But there is now potential for sensors in cities, homes, cars, cell towers and even mobile phones to contribute observational data that could also be fed into forecast models.
Sonia Duarte

UKSA-Business-Plan-April-2019-to-March-2022.pdf - 3 views

  • At this point in time the forecasts for the years beyond 2019/20 are under consideration as we continue to develop our future strategy and bid for the forthcoming Spending Review (2019). As stated previously in the plan 2019/20 is a key year for us in securing the funding required to achieve our ambitions.
  • we will have met our agreed financial targets as part of Spending Review (2015). We also remain broadly on track to deliver our target level of efficiencies over the Spending Review period as indicated in Figure 2 below.
  • Modernising Corporate Support (Efficient) Improving our oversight • Progress with corporate systems improvement projects – milestones in the delivery of automation, workforce planning, improvements to systems controls (new). Delivering value from our resources • Meeting our financial delegations – Budget/forecast accuracy. • Delivering our agreed benefits – Census benefits and ESTP benefits – track of deliveries (new). • Delivering our agreed efficiencies – over the SR15 period.
  • ...2 more annotations...
  • we are nearing the Census delivery date in 2021 with the funding required for Census and Data Collection Transformation activities increasing significantly over the next three financial years.
  • Table 1: High level Authority funding position 2019/20 – based on the Main Supply Estimate agreement with HM Treasury Funding Stream £’m2018/19 2019/20 Variance Resource SR15 Baseline164.85 156.95 (7.89) Income funded24.26 26.60 2.34 Census and Data Collection Transformation94.00 104.00 10.00 Bean HMT Contribution4.00 9.00 5.00 Trade Statistics-2.40 2.40 Pension -4.20 4.20 Budget Cover Transfers0.50 (0.22) (0.72) Subtotal Resource2 8 7.61 302.93 15.33 Capital 13.43 7. 0 0 (6.43) Depreciation 2 3 .10 21.30 (1.80) Income Target24.26 26.60 2.34 Annually Managed Expenditure (AME) (0.84)(1.00)(0.16)
Sonia Duarte

Financials and leadership - Wikimedia Foundation - 0 views

  • 40% Direct support to websites
  • ongoing engineering improvements, product development, design and research, and legal support.
  • 34% Direct support to communities
  • ...5 more annotations...
  • strengthen these communities through grants, programs, events, trainings, partnerships, tools to augment contributor capacity, and support for the legal defense of editors.
  • 12% Fundraising
  • Wikimedia is supported entirely by donations.
  • 14% Administration and governance
  • We manage funds and resources responsibly to recruit and support skilled, passionate staff
diaszasz

Cost Structure Block in Business Model Canvas | Cleverism - 0 views

  •  
    Helpful overview of key things to consider when looking at the cost structure of data institutions
diaszasz

Factual | Business Listings in Factual Data - 1 views

  •  
    Factual started out as an aggregator that allowed organisations to deposit point of interest data to create an aggregated set. Their original business model, IIRC, was around licensing that dataset, but contributors got free access or favourable terms. I've noticed that they've changed their model, so the work of contributing the data is done via "Trusted Data Contributors" who appear to take on the work and responsibility for vetting upstream contributions. https://www.factual.com/updatelisting/ Sharing because I think the evolution is interesting, as is the approach to certifying upstream contributions. Relevant to the certification/audit discussion. Similar issues with some of the alt data ecosystem too I expect.
  •  
    Some background on their early days in this 2012 podcast https://cloudofdata.com/2012/01/data-market-chat-tyler-bell-discusses-factual/
fionntan

Audits, External - Encyclopedia - Business Terms | Inc.com - 1 views

  • The auditor's unqualified report contains three paragraphs. The introductory paragraph identifies the financial statements audited, states that management is responsible for those statements, and asserts that the auditor is responsible for expressing an opinion on them. The scope paragraph describes what the auditor has done and specifically states that the auditor has examined the financial statements in accordance with generally accepted auditing standards and has performed appropriate tests. The opinion paragraph expresses the auditor's opinion (or formally announces his or her lack of opinion and why) on whether the statements are in accordance with generally accepted accounting principles.
  • Major types of audits conducted by external auditors include the financial statements audit, the operational audit, and the compliance audit. A financial statement audit (or attest audit) examines financial statements, records, and related operations to ascertain adherence to generally accepted accounting principles. An operational audit examines an organization's activities in order to assess performances and develop recommendations for improvements, or further action. Auditors perform statutory audits which are performed to comply with the requirements of a governing body, such as a federal, state, or city government or agency. A compliance audit has as its objective the determination of whether an organization is following established procedures or rules.
  • the auditor's final report to management often includes recommendations on methodologies of improving internal controls that are in place.
  • ...1 more annotation...
  • The primary goal of external auditing is to determine the extent to which the organization adheres to managerial policies, procedures, and requirements. The independent or external auditor is not an employee of the organization. He or she performs an examination with the objective of issuing a report containing an opinion on a client's financial statements. The attest function of external auditing refers to the auditor's expression of an opinion on a company's financial statements. The typical independent audit leads to an attestation regarding the fairness and dependability of the statements. This is communicated to the officials of the audited entity in the form of a written report accompanying the statements (an oral presentation of findings may sometimes be requested as well).
fionntan

Evolution of Auditing: From the Traditional Approach to the Future Audit - 3 views

shared by fionntan on 10 Jun 20 - No Cached
  • advances in information technology in conjunction with real-time approaches to conducting business are challenging the auditing profession.
  • emphasis has historically been placed on a periodic, backward-looking approach whereby key events and activities are often identified long after their occurrence or simply undetected. Given that recent developments and technologies facilitated a movement away from the historical paradigm and toward a more proactive approach, it is essential that auditors understand what the future audit entails and how they might begin to envision a logical progression to such a state
  • Furthermore, refinements of audit standards generally consisted of reactionary measures that occurred in response to significant negative business events.
    • fionntan
       
      audits, accounting seem to happen after things have gone wrong and new regulations created
  • ...5 more annotations...
  • As a result, the AICPA issued Statement on Auditing Procedure (SAP) No. 1 in October 1939 and it required that auditors inspect inventories and confirm receivables. Consequently, auditors became responsible for auditing the business entity itself rather than simply relying upon management verification routines.
  • First, in 1961 Felix Kaufman wrote Electronic Data Processing and Auditing. The book compares auditing around and through the computer. Historically, auditing around the computer entails traditional manual procedures in which the existence of automated equipment is ignored. As such, the computer is treated as a black box. In this context, auditors rely upon physical inputs to and outputs from automated devices and do not concern themselves with how processing actually occurs within the system(s). Conversely, auditing through the computer involves actual use of computer systems in testing both controls and transactions. Finally, auditing with the computer entails direct evaluation of computer software, hardware, and processes. Consequently, auditing through the computer or with the computer is able to provide a much higher level of assurance when contrasted with auditing around the computer.
  • Although some aspects of the traditional audit will continue to hold value, the audit of the future provides opportunities to increase the use of automated tools and remains a key for offering improved assurances relative to the responsible management and utilization of stakeholder assets.
  • As previously mentioned, basic CAATS contain capabilities to enhance audit effectiveness and efficiency. However, they do not operate on a 24/7 basis and therefore fail to construct a truly continuous auditing environment whereby exceptions and anomalies may be identified as they occur. Alternatively stated, they do not work with real-time or close to real-time data streams and, thus, are not able to address questionable events such as potential fraud or irregularities in an optimized fashion. Cangemi (2010) argues that, given the recent advances in business technologies, the continuing emphasis on the backward looking audit is simply an outdated philosophy. Instead, he believes that real-time solutions are needed.
  • Future audit approaches would likely require auditors, regulators, and standards setters to make significant adjustments. Such adjustments might include (1) changes in the timing and frequency of the audit, (2) increased education in technology and analytic methods, (3) adoption of full population examination instead of sampling, (4) re-examination of concepts such as materiality and independence, and (5) mandating the provisioning of the audit data standard. Auditors would need to possess substantial technical and analytical skills
Ben Snaith

Patterns of data institution that support people to steward data themselves, or become ... - 0 views

  • it enables people to contribute data about them to it and, on a case-by-case basis, people can choose to permit third parties to access that data. This is the pattern that many personal data stores and personal data management systems adopt in holding data and enabling users to unlock new apps and services that can plug into it. Health Bank enables people to upload their medical records and other information like wearable readings and scans to share with doctors or ‘loved ones’ to help manage their care; Japan’s accredited information banks might undertake a similar role. Other examples — such as Savvy and Datacoup — seem to be focused on sharing data with market research companies willing to offer a form of payment. Some digital identity services may also conform to this pattern.
  • it enables people to contribute data about them to it and, on a case-by-case basis, people can choose whether that data is shared with third parties as part of aggregate datasets. OpenHumans is an example that enables communities of people to share data for group studies and other activities. Owners of a MIDATA account can “actively contribute to medical research and clinical studies by granting selective access to their personal data”. The approach put forward by the European DECODE project would seem to support this type of individual buy-in to collective data sharing, in that case with a civic purpose. The concept of data unions advocated by Streamr seeks to create financial value for individuals by creating aggregate collections of data in this way. Although Salus Coop asks its users to “share and govern [their] data together.. to put it at the service of collective return”, it looks as though individuals can choose which uses to put it to.
  • it enables people to contribute data about them to it and decisions about what third parties can access aggregate datasets are taken collectively. As an example, The Good Data seeks to sell browsing data generated by its users “entirely on their members’ terms… [where] any member can participate in deciding these rules”. The members of the Holland Health Data Cooperative would similarly appear to “determine what happens to their data” collectively, as would drivers and other workers who contribute data about them to Workers Info Exchange.
  • ...6 more annotations...
  • it enables people to contribute data about them and defer authority to it to decide who can access the data. A high-profile proposal of this pattern comes in the form of ‘bottom-up data trusts’ — Mozilla Fellow Anouk Ruhaak has described scenarios where multiple people “hand over their data assets or data rights to a trustee”. Some personal data stores and personal information management systems will also operate under this kind of delegated authority within particular parameters or settings.
  • people entrust it to mediate their relationships with services that collect data about them. This is more related to decisions about data collection rather than decisions about access to existing data, but involves the stewardship of data nonetheless. For example, Tom Steinberg has described a scenario whereby “you would nominate a Personal Data Representative to make choices for you about which apps can do what with your data.. [it] could be a big internet company, it could be a church, it could be a trade union, or it could be a dedicated rights group like the Electronic Frontier Foundation”. Companies like Disconnect.Me and Jumbo are newer examples of this type of approach in practice.
  • it enables people to collect or create new data. Again, this pattern describes the collection rather than the re-use of existing data. For example, OpenBenches enables volunteers to contribute information about memorial benches, and OpenStreetMap does similar at much larger scale to collaboratively create and maintain a free map of the world. The ODI has published research into well-known collaboratively maintained datasets, including Wikidata, Wikipedia and MusicBrainz, and a library of related design patterns. I’ve included this pattern here as to me it represents a way for people to be directly involved in the stewardship of data, personal or not.
  • it collects data in providing a service to users and, on a case-by-case basis, users can share that data directly with third parties. This pattern enables users to unlock new services by sharing data about them (such as via Open Banking and other initiatives labelled as ‘data portability’), or to donate data for broader notions of good (such as Strava’s settings that enable its users to contribute data about them to aggregate datasets shared with cities for planning). I like IF’s catalogue of approaches for enabling people to permit access to data in this way, and its work to show how services can design for the fact that data is often about multiple people.
  • it collects data by providing a service to users and shares that data directly with third parties as provisioned for in its Terms and Conditions. This typically happens when we agree to Ts&Cs that allow data about us to be shared with third parties of an organisation’s choice, such as for advertising, and so might be considered a ‘dark’ pattern. However, some data collectors are beginning to do this for more public, educational or charitable purposes — such as Uber’s sharing of aggregations of data with cities via the SharedStreets initiative. Although the only real involvement we have here in stewarding data is in choosing to use the service, might we not begin to choose between services, in part, based on how well they act as data institutions?
  • I echo the point that Nesta recently made in their paper on ‘citizen-led data governance’, that “while it can be useful to assign labels to different approaches, in reality no clear-cut boundary exists between each of the models, and many of the models may overlap”
‹ Previous 21 - 40 Next ›
Showing 20 items per page