This presentation was given at Data Modeling Zone 2016 in Berlin.
It summarizes the why,how and what of PS-3C, a new proposed way of ensemble based data modeling. This approach could be used for data modelling in a Data Centric Data architecture that is many based on NoSQL (aggregate) based databases, Hive and / or SQL-on-Hadoop / SQL-on-anything solutions.
This document proposes solutions for smart solid waste management in cities. It suggests replacing existing garbage bins with new Wi-Fi enabled sensor bins that can detect fill levels and send signals to collection vehicles. This would optimize collection routes and resources. Data from the smart bins could provide metrics like waste collected over time and bin maintenance needs. The document also discusses applying the 3R methodology of reduce, reuse and recycle to minimize waste. Other proposed solutions include converting waste to energy through waste-to-power plants, using plastic waste to construct roads, and generating biogas from organic waste.
Two hour lecture I gave at the Jyväskylä Summer School. The purpose of the talk is to give a quick non-technical overview of concepts and methodologies in data science. Topics include a wide overview of both pattern mining and machine learning.
See also Part 2 of the lecture: Industrial Data Science. You can find it in my profile (click the face)
Presentation at Data ScienceTech Institute campuses, Paris and Nice, May 2016 , including Intro, Data Science History and Terms; 10 Real-World Data Science Lessons; Data Science Now: Polls & Trends; Data Science Roles; Data Science Job Trends; and Data Science Future
Internet of Things & Its application in Smart AgricultureMohammad Zakriya
As we know Agriculture plays vital role in the development of agricultural country. In India about 70% of population depends upon farming and one third of the nation’s capital comes from farming. Issues concerning agriculture have been always hindering the development of the country. The only solution to this problem is smart agriculture by modernizing the current traditional methods of agriculture. Hence the project aims at making agriculture smart using automation and IoT technologies.
Data Science Tutorial | Introduction To Data Science | Data Science Training ...Edureka!
This Edureka Data Science tutorial will help you understand in and out of Data Science with examples. This tutorial is ideal for both beginners as well as professionals who want to learn or brush up their Data Science concepts. Below are the topics covered in this tutorial:
1. Why Data Science?
2. What is Data Science?
3. Who is a Data Scientist?
4. How a Problem is Solved in Data Science?
5. Data Science Components
- The sixth sense technology allows users to interact with digital information by using hand gestures without any hardware devices. It was first developed in 1990 as a wearable computer and camera system.
- The key components are a camera to track hand gestures, a projector to display information onto surfaces, and a mobile device to handle internet connectivity. The camera sends gesture data to the mobile device for processing using computer vision techniques.
- Applications include using hand gestures to draw on surfaces, get flight information by making circular gestures, and make calls by typing on an projected keypad. The technology aims to seamlessly connect the physical and digital world.
Computer vision is a field that uses techniques to electronically perceive and understand images. It involves acquiring, processing, analyzing and understanding images and can take forms like video sequences. Computer vision aims to duplicate human vision abilities through artificial systems. It has applications in areas like manufacturing inspection, medical imaging, robotics, traffic monitoring and more. Some techniques used in computer vision include image acquisition, preprocessing, feature extraction, detection, recognition and interpretation.
How to Become a Data Scientist
SF Data Science Meetup, June 30, 2014
Video of this talk is available here: https://www.youtube.com/watch?v=c52IOlnPw08
More information at: http://www.zipfianacademy.com
Zipfian Academy @ Crowdflower
This document proposes solutions for smart solid waste management in cities. It suggests replacing existing garbage bins with new Wi-Fi enabled sensor bins that can detect fill levels and send signals to collection vehicles. This would optimize collection routes and resources. Data from the smart bins could provide metrics like waste collected over time and bin maintenance needs. The document also discusses applying the 3R methodology of reduce, reuse and recycle to minimize waste. Other proposed solutions include converting waste to energy through waste-to-power plants, using plastic waste to construct roads, and generating biogas from organic waste.
Two hour lecture I gave at the Jyväskylä Summer School. The purpose of the talk is to give a quick non-technical overview of concepts and methodologies in data science. Topics include a wide overview of both pattern mining and machine learning.
See also Part 2 of the lecture: Industrial Data Science. You can find it in my profile (click the face)
Presentation at Data ScienceTech Institute campuses, Paris and Nice, May 2016 , including Intro, Data Science History and Terms; 10 Real-World Data Science Lessons; Data Science Now: Polls & Trends; Data Science Roles; Data Science Job Trends; and Data Science Future
Internet of Things & Its application in Smart AgricultureMohammad Zakriya
As we know Agriculture plays vital role in the development of agricultural country. In India about 70% of population depends upon farming and one third of the nation’s capital comes from farming. Issues concerning agriculture have been always hindering the development of the country. The only solution to this problem is smart agriculture by modernizing the current traditional methods of agriculture. Hence the project aims at making agriculture smart using automation and IoT technologies.
Data Science Tutorial | Introduction To Data Science | Data Science Training ...Edureka!
This Edureka Data Science tutorial will help you understand in and out of Data Science with examples. This tutorial is ideal for both beginners as well as professionals who want to learn or brush up their Data Science concepts. Below are the topics covered in this tutorial:
1. Why Data Science?
2. What is Data Science?
3. Who is a Data Scientist?
4. How a Problem is Solved in Data Science?
5. Data Science Components
- The sixth sense technology allows users to interact with digital information by using hand gestures without any hardware devices. It was first developed in 1990 as a wearable computer and camera system.
- The key components are a camera to track hand gestures, a projector to display information onto surfaces, and a mobile device to handle internet connectivity. The camera sends gesture data to the mobile device for processing using computer vision techniques.
- Applications include using hand gestures to draw on surfaces, get flight information by making circular gestures, and make calls by typing on an projected keypad. The technology aims to seamlessly connect the physical and digital world.
Computer vision is a field that uses techniques to electronically perceive and understand images. It involves acquiring, processing, analyzing and understanding images and can take forms like video sequences. Computer vision aims to duplicate human vision abilities through artificial systems. It has applications in areas like manufacturing inspection, medical imaging, robotics, traffic monitoring and more. Some techniques used in computer vision include image acquisition, preprocessing, feature extraction, detection, recognition and interpretation.
How to Become a Data Scientist
SF Data Science Meetup, June 30, 2014
Video of this talk is available here: https://www.youtube.com/watch?v=c52IOlnPw08
More information at: http://www.zipfianacademy.com
Zipfian Academy @ Crowdflower
Data Science Training | Data Science Tutorial | Data Science Certification | ...Edureka!
This Edureka Data Science Training will help you understand what is Data Science and you will learn about different Data Science components and concepts. This tutorial is ideal for both beginners as well as professionals who want to learn or brush up their Data Science concepts. Below are the topics covered in this tutorial:
1. What is Data Science?
2. Job Roles in Data Science
3. Components of Data Science
4. Concepts of Statistics
5. Power of Data Visualization
6. Introduction to Machine Learning using R
7. Supervised & Unsupervised Learning
8. Classification, Clustering & Recommenders
9. Text Mining & Time Series
10. Deep Learning
To take a structured training on Data Science, you can check complete details of our Data Science Certification Training course here: https://goo.gl/OCfxP2
This document provides an overview of artificial intelligence and machine learning. It begins by defining AI as computer systems that can perform tasks autonomously and adaptively. Machine learning is described as getting computers to learn without being explicitly programmed. Examples of machine learning in daily life are discussed. The basics of supervised and unsupervised learning are explained. Ethical issues around AI like bias, fairness, and determining appropriate use are then discussed. Options for addressing these issues like ensuring diversity of data and viewpoints are presented. The document concludes by providing recommendations for further learning.
The document outlines a data science roadmap that covers fundamental concepts, statistics, programming, machine learning, text mining, data visualization, big data, data ingestion, data munging, and tools. It provides the percentage of time that should be spent on each topic, and lists specific techniques in each area, such as linear regression, decision trees, and MapReduce in big data.
Artificial Intelligence & Business Application.pptxShamraoGhodake2
The document discusses various topics related to artificial intelligence and knowledge management including:
1) An overview of artificial intelligence and its applications in management such as supply chain optimization.
2) Key concepts in knowledge management such as knowledge creation, capture, sharing and application.
3) Frameworks for knowledge management such as Nonaka's SECI model and the Knowledge Management Maturity Model.
4) The knowledge management cycle and different approaches to knowledge management.
Data Science For Beginners | Who Is A Data Scientist? | Data Science Tutorial...Edureka!
This document outlines an agenda for a data science training presentation. The agenda includes sections on why data science, what data science is, who a data scientist is, what they do, how to solve problems in data science, data science tools, and a demo. Key points are that data science uses tools, algorithms and machine learning to discover patterns in raw data and gain insights. It involves tasks like processing, cleaning, mining and modeling data, as well as communicating results. The problem solving process involves discovery, preparation, planning, building, operationalizing and communicating models.
IoT Based Garbage Monitoring System pptRanjan Gupta
1) A group of students presented on an IOT Garbage Monitoring System to help keep cities clean.
2) The system uses ultrasonic sensors and a microcontroller to monitor garbage levels in bins and displays the status on an LCD screen and web page.
3) When fully implemented, the system will help support initiatives like Swachh Bharat Mission by enabling real-time garbage monitoring and efficient collection.
This Isn't 'Big Data.' It's Just Bad Data.Peter Orszag
With response rates that have declined to under 10 percent, public opinion polls are increasingly unreliable. Perhaps even more concerning, though, is that the same phenomenon is hindering surveys used for official government statistics, including the Current Population Survey, the Survey of Income and Program Participation and the American Community Survey.
A changing market landscape and open source innovations are having a dramatic impact on the consumability and ease of use of data science tools. Join this session to learn about the impact these trends and changes will have on the future of data science. If you are a data scientist, or if your organization relies on cutting edge analytics, you won't want to miss this!
The document discusses mind reading computers. It begins with an introduction explaining that mind reading computers analyze facial expressions and gestures in real time to infer mental states. It then discusses the technology used, including a futuristic headband that measures blood oxygen levels around the brain. Finally, it discusses potential applications of mind reading computers, such as helping communicate with coma patients or allowing people to control devices with their thoughts.
Smart Eye's objective is to be the leading provider of eye tracking systems for vehicles and research. It aims to understand, assist with, and predict human intentions through eye tracking technology. Eye tracking is useful for research into visual attention and consumer purchasing behavior since most information processing and purchases are visually driven. Smart Eye was founded in 1999 and has since released several eye tracking products, becoming a leader in the automotive industry. Advantages include speed, ease of use, and ability to determine areas of interest, while disadvantages include cost and difficulty tracking some users.
The document provides an overview of artificial intelligence (AI), including its history, definition, examples, advantages, and disadvantages. It traces the origins of AI concepts back to ancient Greece and discusses early milestones like the Turing test. Examples of modern AI applications mentioned include Google Maps, facial recognition, chatbots, and automated payments. While AI can reduce human error and perform dangerous tasks, disadvantages include high costs and an inability to think creatively.
The Blue Eye technology aims to give computers human-like perceptual and sensory abilities. It uses sensors to identify a user's actions, extract key information, analyze it, and determine their physical, emotional, and informational state. The Blue Eye system uses a personal area network connecting data acquisition units containing sensors to a central system unit for processing. Sensors like the Jazz Multisensor can track eye movement, blood oxygenation, acceleration, and light intensity. The central system analyzes incoming sensor data and records conclusions. Potential applications include power plant control rooms, ship bridges, and professional drivers, helping avoid human errors from tiredness.
Data Science is a wonderful technology that has applications in almost every field. Let's learn the basics of this domain on 16th March at (time).
Agenda
1. What is Data Science? How is it different from ML, DL, and AI
2. Why is this skill in demand?
3. What are some popular applications of Data Science
4. Popular tools and frameworks used in Data Science
Intelligent Document Processing in Healthcare. Choosing the Right Solutions.Provectus
Healthcare organizations generate piles of documents and forms in different formats, making it difficult to achieve operational excellence and streamline business processes. Manual entry and OCR are no longer viable, and healthcare entities are looking for new solutions to handle documents.
In this presentation you can learn about:
- Healthcare document types and use cases
- IDP framework: building blocks for document processing solutions
- The document processing market landscape
- Methodology for solution evaluation: comparing apples to apples
Whether you are looking for a ready-made solution or plan to build a custom solution of your own, this webinar will help you find the best fit for your healthcare use cases.
This document provides an overview of getting started with data science using Python. It discusses what data science is, why it is in high demand, and the typical skills and backgrounds of data scientists. It then covers popular Python libraries for data science like NumPy, Pandas, Scikit-Learn, TensorFlow, and Keras. Common data science steps are outlined including data gathering, preparation, exploration, model building, validation, and deployment. Example applications and case studies are discussed along with resources for learning including podcasts, websites, communities, books, and TV shows.
This is a project report on Smart Dustbin Using IOT Prepared By Lakshya Pandey, Second Year Electrical Engineering Student of Bipin Tripathi Kumaon Institute of Technology (BTKIT), Dwarahat
All Rights Reserved.
Who is a Data Scientist? | How to become a Data Scientist? | Data Science Cou...Edureka!
** Data Scientist Masters' Program: https://www.edureka.co/masters-program/data-scientist-certification **
This Edureka PPT on "Who is a Data Scientist" will help you understand what a data scientist does, their roles and responsibilities, and what the data science profile is all about. You will also get a glimpse of what kind of salary packages and career opportunities the data science domain offers.
Below topics are covered in this PPT:
Who is a Data Scientist?
What is Data Science?
Who can take up Data Science?
How to become a Data Scientist?
Data Scientist Skills
Data Scientist Roles & Responsibilities
Data Scientist Salary
Follow us to never miss an update in the future.
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
Data Science Training | Data Science Tutorial | Data Science Certification | ...Edureka!
This Edureka Data Science Training will help you understand what is Data Science and you will learn about different Data Science components and concepts. This tutorial is ideal for both beginners as well as professionals who want to learn or brush up their Data Science concepts. Below are the topics covered in this tutorial:
1. What is Data Science?
2. Job Roles in Data Science
3. Components of Data Science
4. Concepts of Statistics
5. Power of Data Visualization
6. Introduction to Machine Learning using R
7. Supervised & Unsupervised Learning
8. Classification, Clustering & Recommenders
9. Text Mining & Time Series
10. Deep Learning
To take a structured training on Data Science, you can check complete details of our Data Science Certification Training course here: https://goo.gl/OCfxP2
This document provides an overview of artificial intelligence and machine learning. It begins by defining AI as computer systems that can perform tasks autonomously and adaptively. Machine learning is described as getting computers to learn without being explicitly programmed. Examples of machine learning in daily life are discussed. The basics of supervised and unsupervised learning are explained. Ethical issues around AI like bias, fairness, and determining appropriate use are then discussed. Options for addressing these issues like ensuring diversity of data and viewpoints are presented. The document concludes by providing recommendations for further learning.
The document outlines a data science roadmap that covers fundamental concepts, statistics, programming, machine learning, text mining, data visualization, big data, data ingestion, data munging, and tools. It provides the percentage of time that should be spent on each topic, and lists specific techniques in each area, such as linear regression, decision trees, and MapReduce in big data.
Artificial Intelligence & Business Application.pptxShamraoGhodake2
The document discusses various topics related to artificial intelligence and knowledge management including:
1) An overview of artificial intelligence and its applications in management such as supply chain optimization.
2) Key concepts in knowledge management such as knowledge creation, capture, sharing and application.
3) Frameworks for knowledge management such as Nonaka's SECI model and the Knowledge Management Maturity Model.
4) The knowledge management cycle and different approaches to knowledge management.
Data Science For Beginners | Who Is A Data Scientist? | Data Science Tutorial...Edureka!
This document outlines an agenda for a data science training presentation. The agenda includes sections on why data science, what data science is, who a data scientist is, what they do, how to solve problems in data science, data science tools, and a demo. Key points are that data science uses tools, algorithms and machine learning to discover patterns in raw data and gain insights. It involves tasks like processing, cleaning, mining and modeling data, as well as communicating results. The problem solving process involves discovery, preparation, planning, building, operationalizing and communicating models.
IoT Based Garbage Monitoring System pptRanjan Gupta
1) A group of students presented on an IOT Garbage Monitoring System to help keep cities clean.
2) The system uses ultrasonic sensors and a microcontroller to monitor garbage levels in bins and displays the status on an LCD screen and web page.
3) When fully implemented, the system will help support initiatives like Swachh Bharat Mission by enabling real-time garbage monitoring and efficient collection.
This Isn't 'Big Data.' It's Just Bad Data.Peter Orszag
With response rates that have declined to under 10 percent, public opinion polls are increasingly unreliable. Perhaps even more concerning, though, is that the same phenomenon is hindering surveys used for official government statistics, including the Current Population Survey, the Survey of Income and Program Participation and the American Community Survey.
A changing market landscape and open source innovations are having a dramatic impact on the consumability and ease of use of data science tools. Join this session to learn about the impact these trends and changes will have on the future of data science. If you are a data scientist, or if your organization relies on cutting edge analytics, you won't want to miss this!
The document discusses mind reading computers. It begins with an introduction explaining that mind reading computers analyze facial expressions and gestures in real time to infer mental states. It then discusses the technology used, including a futuristic headband that measures blood oxygen levels around the brain. Finally, it discusses potential applications of mind reading computers, such as helping communicate with coma patients or allowing people to control devices with their thoughts.
Smart Eye's objective is to be the leading provider of eye tracking systems for vehicles and research. It aims to understand, assist with, and predict human intentions through eye tracking technology. Eye tracking is useful for research into visual attention and consumer purchasing behavior since most information processing and purchases are visually driven. Smart Eye was founded in 1999 and has since released several eye tracking products, becoming a leader in the automotive industry. Advantages include speed, ease of use, and ability to determine areas of interest, while disadvantages include cost and difficulty tracking some users.
The document provides an overview of artificial intelligence (AI), including its history, definition, examples, advantages, and disadvantages. It traces the origins of AI concepts back to ancient Greece and discusses early milestones like the Turing test. Examples of modern AI applications mentioned include Google Maps, facial recognition, chatbots, and automated payments. While AI can reduce human error and perform dangerous tasks, disadvantages include high costs and an inability to think creatively.
The Blue Eye technology aims to give computers human-like perceptual and sensory abilities. It uses sensors to identify a user's actions, extract key information, analyze it, and determine their physical, emotional, and informational state. The Blue Eye system uses a personal area network connecting data acquisition units containing sensors to a central system unit for processing. Sensors like the Jazz Multisensor can track eye movement, blood oxygenation, acceleration, and light intensity. The central system analyzes incoming sensor data and records conclusions. Potential applications include power plant control rooms, ship bridges, and professional drivers, helping avoid human errors from tiredness.
Data Science is a wonderful technology that has applications in almost every field. Let's learn the basics of this domain on 16th March at (time).
Agenda
1. What is Data Science? How is it different from ML, DL, and AI
2. Why is this skill in demand?
3. What are some popular applications of Data Science
4. Popular tools and frameworks used in Data Science
Intelligent Document Processing in Healthcare. Choosing the Right Solutions.Provectus
Healthcare organizations generate piles of documents and forms in different formats, making it difficult to achieve operational excellence and streamline business processes. Manual entry and OCR are no longer viable, and healthcare entities are looking for new solutions to handle documents.
In this presentation you can learn about:
- Healthcare document types and use cases
- IDP framework: building blocks for document processing solutions
- The document processing market landscape
- Methodology for solution evaluation: comparing apples to apples
Whether you are looking for a ready-made solution or plan to build a custom solution of your own, this webinar will help you find the best fit for your healthcare use cases.
This document provides an overview of getting started with data science using Python. It discusses what data science is, why it is in high demand, and the typical skills and backgrounds of data scientists. It then covers popular Python libraries for data science like NumPy, Pandas, Scikit-Learn, TensorFlow, and Keras. Common data science steps are outlined including data gathering, preparation, exploration, model building, validation, and deployment. Example applications and case studies are discussed along with resources for learning including podcasts, websites, communities, books, and TV shows.
This is a project report on Smart Dustbin Using IOT Prepared By Lakshya Pandey, Second Year Electrical Engineering Student of Bipin Tripathi Kumaon Institute of Technology (BTKIT), Dwarahat
All Rights Reserved.
Who is a Data Scientist? | How to become a Data Scientist? | Data Science Cou...Edureka!
** Data Scientist Masters' Program: https://www.edureka.co/masters-program/data-scientist-certification **
This Edureka PPT on "Who is a Data Scientist" will help you understand what a data scientist does, their roles and responsibilities, and what the data science profile is all about. You will also get a glimpse of what kind of salary packages and career opportunities the data science domain offers.
Below topics are covered in this PPT:
Who is a Data Scientist?
What is Data Science?
Who can take up Data Science?
How to become a Data Scientist?
Data Scientist Skills
Data Scientist Roles & Responsibilities
Data Scientist Salary
Follow us to never miss an update in the future.
Instagram: https://www.instagram.com/edureka_learning/
Facebook: https://www.facebook.com/edurekaIN/
Twitter: https://twitter.com/edurekain
LinkedIn: https://www.linkedin.com/company/edureka
This presentation describes a multi-faceted project used in an Intermediate Accounting class where students interact with accounting professionals. How this project contributes to student success as well as how such a project aligns with ACBSP standards are discussed.
Dear Sir/Madam,
I am writing to express my strong interest as a Civil Draftsman for the position recently posted on your website. I did civil Draftsman from Government Technical College, Sahiwal, Pakistan. I had 3 years of experience as a draftsman including site supervision.
I worked in Continental Overseas Construction Services, Lahore, Pakistan as Site Supervisor (Infrastructure work and execution work). I am currently working as an Architectural Draftsman at Aamer Fayyaz & Associates, Lahore, Pakistan. As you will see from my resume, I have the requisite skills you are looking for to be successful in this position. More importantly, I have the passion and drive the relevant experience in the field of architecture and civil engineering with the well renowned companies like Continental Overseas Construction Services and Aamer fayyaz & Associates, Lahore, Pakistan. Combined with this passion and my knowledge, I am very comfortable with site execution, drawing preparation and report writing. It will be worth writing to mention as I worked with Aamer Fayyaz & Associates and some projects in Lahore before joining them, so it’s not new for me and is also not time consuming to understating the working environment here. I can be reached by phone or email, both listed on my resume. Thank you for your time and consideration; I look forward to speak with you soon.
Sincerely,
Amer Hasan
+92 300 5510084
Over four years, Walsh College targeted improvements to student performance in written communication skills through its MBA capstone course. Assessment of student work identified gaps in sources/evidence and formatting between online and on-ground students. The college addressed these issues by revising assignments to better align with competencies, providing targeted instructional content, and implementing formative writing activities across modalities. These actions closed performance gaps and increased student scores by over 50% on sources/evidence and 13% on formatting from 2011-2014. Walsh College plans to continue refining rubrics, assignments, and online resources to further enhance written communication skills through continuous assessment and improvement cycles.
This document is a resume for Michelle K. Clayton. It summarizes her 20 years of experience in sales and sales training for academic publishing companies, including her roles leading high performing sales teams and consistently exceeding sales goals. She has extensive experience selling and implementing digital learning solutions and is skilled in consultative selling, relationship building, and coaching others.
This document provides guidance on conducting a successful self-study for accreditation. It recommends that the core team be creative, have strong writing and attention to detail skills, and work well under pressure and in teams. Support from campus presidents, faculty, deans, and administration is important. The self-study process involves drafting a narrative and collecting documentation in various areas like faculty information, courses, and assessments. Clear labeling, consistent application of standards, and visual representations should be used. Thorough documentation is key. Timelines should allow more time than anticipated. The completed self-study should answer all parts of the standards clearly and have support for its conclusions. Being prepared for the on-site visit is also advised.
The document discusses how the Baker School of Business and Technology at the Fashion Institute of Technology transformed its institution to fully accept and embrace accreditation from the Accreditation Council for Business Schools and Programs (ACBSP). It outlines the school's history and enrollment growth. In 2012, a strategic review recommended pursuing ACBSP accreditation to streamline curriculum and expand course offerings. Some faculty were initially skeptical but two core approaches - "pull" involving deep faculty involvement and "push" through aggressive process upgrades and communication - helped address concerns and gain support. With faculty-led leadership and project management, as well as multi-level communication through events and updates, the school was able to successfully transform its culture and earn
Capella University uses a collaborative model to support learner success through accreditation. It has a centralized accreditation department that oversees the university's regional and specialized accreditations. Capella's competency-based and professionally aligned programs are assessed using a system aligned with academic and professional standards, with learners tracking their progress through course competency maps. The accreditation cycle involves initiation, discovery, verification, decision making, and maintenance steps to maintain accreditations like ACBSP for business programs.
The document discusses Walsh College's process for simultaneously assessing and grading student work to collect internal assessment data. Walsh College maps assignment instructions and rubric criteria to its core competencies of problem solving and written communication. Faculty integrate functional rubrics into graded assignments in target courses to assess competencies. This allows Walsh to leverage assignments for both grading students and assessing competency achievement. The approach has been implemented in 17 courses so far. Challenges include ensuring faculty assess competencies objectively when grading. Suggestions for overcoming challenges include increasing faculty involvement and periodically assessing results to improve processes.
Top 8 executive administrator resume sampleskerrojom
The document provides resources for executive administrator resumes, cover letters, and interview preparation materials from the website resume123.org. It lists 8 resume samples, tips for writing effective resumes and cover letters, and over 60 interview questions, tips, and other materials to help prepare for an executive administrator interview.
The document describes PSG Institute of Management's efforts to foster entrepreneurship among its students in India. It discusses how the school integrates experiential learning approaches into its curriculum through business plan competitions and mentorship opportunities. Several student-run businesses and social ventures have launched as a result of this focus on hands-on learning. The school aims to develop entrepreneurial skills and kindle the entrepreneurial spirit of its graduates.
This document summarizes an experiential learning class at New England College that partners with an organization called TIST in Kenya. The class includes a 10-day trip to Kenya to work on sustainable development projects. Key components of the class include analyzing the triple bottom line approach, impact analysis methods, adaptive leadership, and developing a sense of global citizenship. Outcomes include conducting a social return on investment analysis of TIST and developing marketing materials to support their carbon offset program. Student feedback praised the unique learning experience and change in perspective gained from the class.
Pass chapter meeting dec 2013 - compression a hidden gem for io heavy databas...Charley Hanania
Compression: a hidden Gem for IO heavy Databases
The limiting factor in most database systems is the ability to read and write data to the IO subsystem.
We're still using storage layouts and methodologies in SQL Server that are a reflection of old spinning media in times gone by.
Until major changes are made to the internal storage layouts, we have "some" hope with options such as data compression, sparse columns and filtered indexes, which not only save space on disk, but also reflect a saving in memory.
In this session we will go over the IO savings technologies presented in SQL Server, and discuss how implementing some of these will assist in your operational performance goals.
Presenter: Charley Hanania, MVP
Charley is Principal Consultant at QS2 AG in Switzerland and has consulted to organisations of all sizes during his extensive career in Database and Platform Consulting.
He's been focussed on SQL Server since v4.2 on OS/2 and with over 15 years of experience in IT he's supported companies in the areas of DB training, development, architecture & administration throughout Europe, America & Australasia.
Communities are Charley's passion and he became active in database communities in the mid 90's, participating in heterogeneous database user groups in Australia. He continues to lead an active role through community events such as Database Days, the European PASS Conference, PASS & the Swiss PASS Chapter.
This presentation will be useful to those who would like to get acquainted with Apache Spark architecture, top features and see some of them in action, e.g. RDD transformations and actions, Spark SQL, etc. Also it covers real life use cases related to one of ours commercial projects and recall roadmap how we’ve integrated Apache Spark into it.
Was presented on Morning@Lohika tech talks in Lviv.
Design by Yarko Filevych: http://www.filevych.com/
The document discusses SQL vs NoSQL databases. It provides background on the proliferation of NoSQL databases and their advantages over relational databases for handling unstructured data, high scalability, and easy distribution. However, it argues that SQL remains well-suited for analytical queries due to its portability, wide use, and the fact that many reporting tools are built for it. The document also presents a case study of how the online gaming company King uses a hybrid of SQL and NoSQL technologies to handle their massive scale of user data and high-volume analytics needs.
Eugene Polonichko "Architecture of modern data warehouse"Lviv Startup Club
The document discusses the architecture of a modern data warehouse using Microsoft technologies. It describes traditional data warehousing approaches and outlines ten characteristics of a modern data warehouse. It then details Microsoft's approach using Azure Data Factory to ingest diverse data types into Azure Blob Storage, Azure Databricks for analytics and data transformation, and Azure SQL Data Warehouse for combined structured data. It also discusses technologies for storage, visualization, and links for further information.
The document summarizes information about a conference on relational database options in Azure. It provides an agenda that includes introductions to Azure SQL Database, SQL Data Warehouse, Azure CosmosDB, and other database platforms on Azure. It also discusses infrastructure as a service (IaaS), platform as a service (PaaS), and software as a service (SaaS) models and options for migrating databases to Azure. The presentation aims to help attendees learn about relational database options in Microsoft's cloud computing platform.
Why you should be mining your data and how to actually do it. Every company needs a rock star. We want it to be you. This session will give real world examples of data mining successes as well as walk you through how to get started down the path of data enlightenment, so that you too can say "I Am A Data Miner℠".
This document provides an overview of Azure SQL Data Warehouse (SQL DWH), a cloud data warehouse service. It discusses SQL DWH's massively parallel processing (MPP) architecture that allows independent scaling of compute and storage. The document demonstrates how to create a SQL DWH, load data using PolyBase, and use common tools. It is intended to help users understand what SQL DWH is, how it works, and common scenarios it can be used for, such as processing large volumes of data without needing to purchase and manage hardware.
Pablo Pazos Gutiérrez gave a talk on developing openEHR systems. He discussed storing openEHR data using different database types, openEHR system architectures that have evolved to be more distributed and service-oriented, generating user interfaces from archetypes and templates, performing archetype-based validation on entered data, querying and visualizing openEHR data, and implementing openEHR over the past 8 years in Latin America.
This document provides an overview of NoSQL and MongoDB. It discusses trends driving the adoption of NoSQL databases like increasing data sizes, more connectedness, and individualization. It covers the different types of NoSQL databases and MongoDB in particular. Key concepts discussed include the CAP theorem, MongoDB's document-oriented data model, and basic CRUD operations in MongoDB using the shell.
Exploring OrientDB as Graph Database model as NoSQL database.
The main goal of this project is to provide theoretical, technical details and debates on some powerful features of OrientDB. We provide some comparison attempts between OrientDB 2.1.8 and SQL Server 2012, they are mostly focused on MovieLens dataset and build recommendation engine.
Secrets of Enterprise Data Mining: SQL Saturday 328 Birmingham ALMark Tabladillo
This document discusses secrets of enterprise data mining. It begins by defining data mining as the automated or semi-automated process of discovering patterns in data. It then discusses how data mining can be applied in various industries like telecommunications, oil and gas, and Volkswagen Group. Finally, it discusses how Microsoft offers solutions for enterprise data mining through SQL Server Analysis Services and Microsoft Azure Machine Learning.
This document outlines several case-based scenarios for demonstrating data science activities using Azure services. Six cases are described:
1) A playground for citizen data scientists to gain an end-to-end understanding of the data science process using a simple UI.
2) Using SQL databases and services for machine learning tasks when all data resides in SQL.
3) Parallel training of models on multiple datasets to automate and scale the training process.
4) Using GPU-enabled environments for training deep learning models requiring GPU acceleration.
5) Leveraging high-speed data processing services when working with large datasets over 1GB.
6) A basic sandbox environment for data scientists, engineers, and analysts providing pre-
Splice Machine's use of Apache Spark and MLflowDatabricks
Splice Machine is an ANSI-SQL Relational Database Management System (RDBMS) on Apache Spark. It has proven low-latency transactional processing (OLTP) as well as analytical processing (OLAP) at petabyte scale. It uses Spark for all analytical computations and leverages HBase for persistence. This talk highlights a new Native Spark Datasource - which enables seamless data movement between Spark Data Frames and Splice Machine tables without serialization and deserialization. This Spark Datasource makes machine learning libraries such as MLlib native to the Splice RDBMS . Splice Machine has now integrated MLflow into its data platform, creating a flexible Data Science Workbench with an RDBMS at its core. The transactional capabilities of Splice Machine integrated with the plethora of DataFrame-compatible libraries and MLflow capabilities manages a complete, real-time workflow of data-to-insights-to-action. In this presentation we will demonstrate Splice Machine's Data Science Workbench and how it leverages Spark and MLflow to create powerful, full-cycle machine learning capabilities on an integrated platform, from transactional updates to data wrangling, experimentation, and deployment, and back again.
The document provides a summary of Mark Hargraves' work experience in business intelligence. It details several roles he held as a senior BI consultant where he developed ETL processes and data warehouses using SQL Server and the Microsoft BI stack. For each role, he extracted data from various source systems, modeled it using a schema called Spider Schema, and built cubes and reports in SSAS and SSRS. The roles showed his experience in full BI project development and remote work for clients in different industries.
The Machine Learning behind the Autonomous Database- EMEA Tour Oct 2019 Sandesh Rao
Autonomous Database is one of the hottest Oracle products where we have attempted to use Machine Learning for several aspects of the service. We take a view on our current state of ML in the Autonomous Database Cloud and how do we process this data in ADW/ATP with zeppelin notebooks to find anomalies in them to troubleshoot them at a scale of several petabytes a year and conduct AIOps. We will cover some sample notebooks to some use cases we will cover are a Log Anomaly timeline which we reduce significant amounts of logs using semi-supervised machine learning techniques to reduce logs and match them in near real time. Some of the other use cases is to use convolution filters to determine maintenance windows within the database workloads , determine best times to do database backups , security anomaly timelines and many others. This presentation will accompany several examples with how to apply these techniques , machine learning knowledge is preferred but not a prerequisite
This document provides an overview of NoSQL databases in Azure. It discusses 7 different database types - key-value, column family, document, graph and Hadoop. For each database type it provides information on what it is, examples of use cases, and how to query or model data. It encourages attendees to explore these databases and stresses that choosing the right database for the job is important.
Strata 2014: Design Challenges for Real Predictive Platforms Max Gasner
The first databases were tightly coupled to their implementation details and use cases, until the relational revolution opened up the field and made database systems flexible enough to support a wide variety of applications with minimal configuration. What will it take to make predictive systems as ubiquitous and easy to use as databases? We’ll discuss the crucial design criteria for future predictive platforms and the kinds of interfaces they need to be able to support, as well as the challenges that lie between the state of the art and the future we envision.
Open Data Science Conference Big Data Infrastructure – Introduction to Hadoop...DataKitchen
The main objective of this workshop is to give the audience hands on experience with several Hadoop technologies and jump start their hadoop journey. In this workshop, you will load data and submit queries using Hadoop! Before jumping in to the technology, the Founders of DataKitchen review Hadoop and some of its technologies (MapReduce, Hive, Pig, Impala and Spark), look at performance, and present a rubric for choosing which technology to use when.
NOTE: To complete hands on poriton in the time allotted, attendees should come with a newly created AWS (Amazon Web Services) Account and complete the other prerequisites found in the DataKitchen blog <http: />.
This presentation is about health care analysis using sentiment analysis .
*this is very useful to students who are doing project on sentiment analysis
*
Discovering Digital Process Twins for What-if Analysis: a Process Mining Appr...Marlon Dumas
This webinar discusses the limitations of traditional approaches for business process simulation based on had-crafted model with restrictive assumptions. It shows how process mining techniques can be assembled together to discover high-fidelity digital twins of end-to-end processes from event data.
Did you know that drowning is a leading cause of unintentional death among young children? According to recent data, children aged 1-4 years are at the highest risk. Let's raise awareness and take steps to prevent these tragic incidents. Supervision, barriers around pools, and learning CPR can make a difference. Stay safe this summer!
Do People Really Know Their Fertility Intentions? Correspondence between Sel...Xiao Xu
Fertility intention data from surveys often serve as a crucial component in modeling fertility behaviors. Yet, the persistent gap between stated intentions and actual fertility decisions, coupled with the prevalence of uncertain responses, has cast doubt on the overall utility of intentions and sparked controversies about their nature. In this study, we use survey data from a representative sample of Dutch women. With the help of open-ended questions (OEQs) on fertility and Natural Language Processing (NLP) methods, we are able to conduct an in-depth analysis of fertility narratives. Specifically, we annotate the (expert) perceived fertility intentions of respondents and compare them to their self-reported intentions from the survey. Through this analysis, we aim to reveal the disparities between self-reported intentions and the narratives. Furthermore, by applying neural topic modeling methods, we could uncover which topics and characteristics are more prevalent among respondents who exhibit a significant discrepancy between their stated intentions and their probable future behavior, as reflected in their narratives.
Optimizing Feldera: Integrating Advanced UDFs and Enhanced SQL Functionality ...mparmparousiskostas
This report explores our contributions to the Feldera Continuous Analytics Platform, aimed at enhancing its real-time data processing capabilities. Our primary advancements include the integration of advanced User-Defined Functions (UDFs) and the enhancement of SQL functionality. Specifically, we introduced Rust-based UDFs for high-performance data transformations and extended SQL to support inline table queries and aggregate functions within INSERT INTO statements. These developments significantly improve Feldera’s ability to handle complex data manipulations and transformations, making it a more versatile and powerful tool for real-time analytics. Through these enhancements, Feldera is now better equipped to support sophisticated continuous data processing needs, enabling users to execute complex analytics with greater efficiency and flexibility.
9. Not Build for
BiG data lake
/ Data CentrICITY
Photo credit: Lake Public Domain, http://www.writeups.org/star-trek-brent-spiner-data/
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
10. ‘Data may first be stored in a
data lake so that it can be explored, cleaned, and prepared.
If it can be structured in a relational format (basically rows and columns)
and needs to be used frequently and kept highly secure, it may go into a
data warehouse.
If it stops being used frequently, it may go back to a HDFS
(Hadoop Distributed File System)-based archive.’
Data Centric / data first ??
THOMAS H. DAVENPORT, WALL STEET JOURNAL OF 3-6-2015
http://blogs.wsj.com/cio/2015/06/03/the-shift-to-a-new-data-architecture/ @rwerschkull
nl.linkedin.com/in/rogierwerschkull
12. Data Flood
Photo credit: Kurayba (https://www.flickr.com/photos/48503330@N08/28564454666/ )
under cc licence (https://creativecommons.org/licenses/by-sa/2.0/)
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
13. The possible result...…
Photo credit: https://highfiveexports.wordpress.com/2010/06/25/3000-pieces-lego-mix-specialty-pieces-rare-pieces-bricks-blocks-
parts-more-ultimate-lot-of-lego-parts-pieces-lego-for-sale-lego-batman-lego-starwars-lego-technic-lego-minifigur/
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
14. But isN’t Data Vault v2
‘made for
Bigdata centric
systems?’
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
15. In
DV2
you still
do this
in one go
Subject
Oriented
Integrated
Time Variant Non-Volatile
EDW
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
16. Coding =
A lot like modelling?
Being Data Centric
conflicts
with the
complex Data
MODELLING
work
http://xkcd.com/844/
24. a) HASHING OF Business keys
Rolling
Stock Nr Datetime Sensor Id Value Concatenated Business Key
Key
Len MD5 Hash
Key
Len
8739
2015-01-22
01:34:27 72A1_FINV 123 8739|2015-01-22 01:34:27|72A1_FINV 34 86ae4c6b0e2e2d5a13a0d11440529aeb 32
8739
2015-01-22
01:34:27 72A1_SLDET 100 8739|2015-01-22 01:34:27|72A1_SLDET 35 51ce9bc292eef407bd7c91a52eebcf2e 32
8739
2015-01-22
01:34:32 72A1_FINV 126 8739|2015-01-22 01:34:32|72A1_FINV 34 9482a41c1fecc4c64b8c437af6cc85e8 32
8739
2015-01-22
01:34:32
13A8_MW_UB
AT_VT 5 8739|2015-01-22 01:34:32|13A8_MW_UBAT_VT 42 e4160914ee55ce0b93f87b23366a0ce3 32
8674
2015-01-22
01:34:26 72A1_FINV 6 8674|2015-01-22 01:34:26|72A1_FINV 34 fcb3e7c8c91e44ce396d908a4948ca65 32
8674
2015-01-22
01:34:26
16A1_HSVER
OND 7 8674|2015-01-22 01:34:26|16A1_HSVEROND 38 fe9098c8c291ad56af5c8afae5169196 32
25. Loses statistical Information
regarding the data distribution
Query optimizers do not like this…
Column family, Document and Key-value databases need a
good (natural) sharding key for (partial) key-
lookups!*
Hashing...……
* http://www.ebaytechblog.com/2012/08/14/cassandra-data-modeling-best-practices-part-2/
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
26. Surrogates keys require
centralized coordination
…and thus can impact the overall system’s scalability and
availability.
A lot of MPP / NoSQL databases simply do not have
them…
B) Surrogate BuSINESS keys
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
28. ‘In my opinion the answer lies in the adoption of the
persistent (Historical) Staging Area concept
(also known as Historical Staging or the History Area).
This basically adopts the fundamentals of a Data Warehouse’
‘The Historical Staging Area effectively ‘acts’ as
Data Lake,
but in a better defined form as data deltas and
event date/times are taken into account.’
36. Identify source / event stream Primary or Unique Key
Use source metadata for this!
Automate the building of a PS ‘around this key’
Take all columns!
Historize using SCD-2 approach
Persistent Staging - how
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
38. Column level
[Load Date Timestamp]
[Load End Date Timestamp]
[Deleted Flag] OR delete as new record
[Source system] on table / file level (lowest possible)
Load End Date Timestamp : possible but difficult…
Requires updates!
Persistent Staging Metadata-2
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
39. ACID is possible in HIVE!
ACID Makes Updates possible
By registering updates as ‘new data’
Reconciliation / compacting when idle / at user command
Use ORC files!!!
PLUS changing the HIVE configuration…
UPDATES IN HIVE? (iSN’t HDFS APPEND ONLY?)
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
40. Hive
Put semi structured data = variable columns in MAP data type
OR use Data storage type that supports schema-evolution:
AVRO, (ORC in development)
Or HBASE…
It only has one data type (byte), schema is ‘applied’
Schema can be different for every row
What about SEMI-STRUCTURED Data?
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
43. Always starts with Conceptual data modeling
NOT the primary location of Data & History
Virtualised (only if performance allows). Should be deterministic!
No Link Satellites
No Surrogate or Hash Keys, only ‘Contatenated Natural Business Keys’
Explicit Helper entities
Like Data Vault(2) BUT...
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
45. a UNIQUE, Domain specific point of integration
…a business entity
…within it’s own domain
…does not necessarily need to be Enterprise Wide!
Business Concept (BC)
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
46. Why not ‘enterprise wide?’
Company
Customer
Sales
Customer
International
Sales
Customer
Local
Sales
Customer
Marketing
Customer
Customer …
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
47. Entity level
[Description]
[Owner / Responsible]
Column level
[Load Date Timestamp]
[Source system] on table / file level (lowest possible)
Business Concept Metadata
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
49. Most easy entity to be virtualised
(if performance allows)
No Hashing & No surrogate
BUSINESSKEYS!
(not by default at least!)
BC: important notes
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
51. Containts Context about a Concept
In a historical way
…Like a Data Vault Satellite
Every CC belongs to only one BC
Seperate entity per source system / table / stream
Concept Context (CC)
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
52. Entity level
[Description]
[Owner / Responsible]
Column level
[Load Date Timestamp]
[Source system] on table / file level (lowest possible)
Not mandatory for streaming data:
• [Load End Date Timestamp]
• [Deleted Flag] OR register a delete as a new record
Concept Context Metadata
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
53. Example
NSR-Station
NS-
Travelcard
NS-
Trainseries
NS-
Traveller
[valuation]
Source: NSS2 table p
[description]
Source:NSS1 table x
[adres]
Source: NSS1 table y
[description]
Source: NTR table q
[ovchip_
personal]
Source: NSR table r
[ovchip_
on-usage
Source: NSR table s
[personal_
details]
Source: NSR table t
[adres_
data]
Source: NSR table t @rwerschkull
nl.linkedin.com/in/rogierwerschkull
55. More difficult to be virtualised
Depends on semantic gap with source!
But do make virtual when ‘streaming data’ is necessary!
Because we have PS layer
Exposing all columns not necessary!
Refactoring is more easy…
BC: important notes
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
57. Relations between Concepts
+ Context
In a historical way
…Merger of Data Vault Link + Link Satellite
Must ALWAYS have a driving key defined
= a (sub)set of keys that make a Connector unique at one point
in time
Connector (C)
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
58. Explicitly defining a driving key as metadata…
Gives business understanding!
Makes it possible Connector can correctly handle delta data
deliveries…
• so that a change (on the driving key)
• is not registered as a new ‘connection’
Connector Driving key
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
59. Entity level
[Description]
[Owner / Responsible]
Column level
[Load Date Timestamp]
[Source system] on table / file level (lowest possible)
Not mandatory for streaming data:
• [Load End Date Timestamp]
• [Deleted Flag] OR register a delete as a new record
Connector Metadata
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
60. Example
NSR-Station
NS-
Travelcard
NS-
Trainseries
NS-
Traveller
[valuation]
Source: NSS2 table p
[description]
Source:NSS1 table x
[adres]
Source: NSS1 table y
[description]
Source: NTR table q
[ovchip_
personal]
Source: NSR table r
[ovchip_
on-usage
Source: NSR table s
[personal_
details]
Source: NSR table t
[adres_
data]
Source: NSR table t @rwerschkull
nl.linkedin.com/in/rogierwerschkull
NSR-
Travelmovement
Checkin timestamp
from
to
Driving Key:
NS-Travelcard
+Checkin timestamp
65. To help switching from sources that are Tied together
by technical (surrogate) keys…
To a Business Key based model
It’s a LOOKUP table that translates the technical to the
Business Key
Business Alias
66. Example
NSR-Station
[valuation]
Source: NSS2 table p
[description]
Source:NSS1 table x[adres]
Source: NSS1 table y
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
BA-NSS1
BA-NSS2
Key Lookup voor NSS1 source tables
Key Lookup voor NSS2 source tables
68. Has a 1 on (0,1) relation with a Business Concept
More difficult to be virtualised
Lookup table should be kept small!
Therefore: DO NOT do key lookup in Concept Context entity!
Load / generate together with BC
Preferably ‘in memory’ somehow…
BA: important Details
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
70. Integrate the validity timelines of Concept Contexts
belonging to a Business Concept
Like a Data Vault Point-in-time construct
But Mandatory!
And with a clearly defined and performant
approach!
BC-Timeline
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
71. Example
NSR-Station
[valuation]
Source: NSS2 table p
[adres]
Source: NSS1 table y
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
BK_NSR-
Station
WOZ
waarde
Waarde
Ratingbureau X
META_Laad_dts META_Laad_eind_dts
Ut 20 milj 18 milj 1-1-2014 22:00:00 1-1-2015 21:59:59
Ut 22 milj 18 milj 1-1-2015 22:00:00 1-3-2016 21:59:59
Ut 22 milj 23 milj 1-3-2016 22:00:00 31-12-9999 00:00:00
BK_NSR-
Station
Combined_Load_dts Combined_Load_end_dts
Ut 5-6-2013 22:00:00 1-1-2014 21:59:59
Ut 1-1-2014 22:00:00 1-1-2015 21:59:59
Ut 1-1-2015 22:00:00 4-7-2015 21:59:59
Ut 4-7-2015 22:00:00 1-3-2016 21:59:59
Ut 1-3-2016 22:00:00 31-12-9999 00:00:00
Asd 5-6-2013 22:00:00 31-12-9999 00:00:00
BK_NSR-
Station
Postadres_
postcode
GPS … META_
source
META_Load_dts META_Load_end_dts
Ut 3500GJ 52.08954, 5.11064 … NSS1_y 5-6-2013 22:00:00 4-7-2015 21:59:59
Ut 3511 CE 52.37269, 4.89299 … NSS1_y 4-7-2015 22:00:00 31-12-9999 00:00:00
Asd 1012 AB 52.37269, 4.89299 … NSS1_y 5-6-2013 22:00:00 31-12-9999 00:00:00
Asa … … … … … …
BCT
73. What
Makes PS-3C
a Different Ensemble?
Business
Concept
X
Concept
Context
X-A
Concept
Context
X-B
Business
Concept
Y
Concept
Context
Y-A
Concept
Context
Y-B
Concept
Context
Y-C
Connector
Business
Alias A
Business
Alias B
BC-Timeline
X
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
74. 1) Explicitly Splitting The work
Data
+
History
Subjects
+
Integration
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
75. 2) NO HASHEDBUSINeSS KEYS
...or surrogate keys
http://www.cannabisculture.com/files/images/6/hashbrick.JPG
Only
Concatenated
ones
@rwerschkull
nl.linkedin.com/in/rogierwerschkull
76. 3) Less
joins
Relation
+Technical validity timeline
+ Relation context
Together in one entity
Photo credit: Public Domain
@rwerschkull
nl.linkedin.com/in/rogierwerschkull