SQLBits Module 2 RStats Introduction to R and Statistics. This is a 90 minute segment of a full preconference workshop, focusing on data analytics with R.
Digital Pragmatism with Business Intelligence, Big Data and Data VisualisationJen Stirrup
Contact details:
[email protected]
In a world where the HiPPO’s (Highest Paid Person’s Opinion) is final, how can we use technology to drive the organisation towards data-driven decision making as part of their organizational DNA? R provides a range of functionality in machine learning, but we need to expose its richness in a world where it is made accessible to decision makers. Using Data Storytelling with R, we can imprint data in the culture of the organization by making it easily accessible to everyone, including decision makers. Together, the insights and process of machine learning are combined with data visualisation to help organisations derive value and insights from big and little data.
Business Intelligence Barista: What DataViz Tool to Use, and When?Jen Stirrup
Choosing a data visualization tool is like being a barista serving coffee: everyone wants their data, their way, personalized, fast, and perfect. Many organizations have a cottage industry of data visualization tools, and it's difficult to know what tool to use, and when. Different tools exist in different departments, and if it doesn't meet the user requirements, the default position is to go back to Excel and move the data around there.
This session will examine data visualization tools such as SSRS Excel, Tableau, QlikView, Datazen, Kibana and PowerBI, in order to craft and blend your data visualization tools to serve your data customers better.
NoSQL and Data Modeling for Data ModelersKaren Lopez
Karen Lopez's presentation for data modelers and data architects. Why data modeling is still relevant for big data and NoSQL projects.
Plus 10 tips for data modelers for working on NoSQL projects.
Sql rally amsterdam Aanalysing data with Power BI and HiveJen Stirrup
Analyzing Data with Power View (Level 100)
Jen Stirrup
Come learn about the best ways to present data to your Business Intelligence data consumers, and see how to apply these principles in Power View, Microsoft's data visualization tool. Using demos, we will investigate Power View based on current cognitive research around data visualization principles from such experts as Stephen Few, Edware Tufte, and others. We will then examine how data can be analyzed with Power View and look at where Power View is supplemented by other parts of the Microsoft Business Intelligence stack.
Data visualization in data science: exploratory EDA, explanatory. Anscobe's quartet, design principles, visual encoding, design engineering and journalism, choosing the right graph, narrative structures, technology and tools.
This document discusses appropriate and inappropriate use cases for Apache Spark based on the type of data and workload. It provides examples of good uses, such as batch processing, ETL, and machine learning/data science. It also gives examples of bad uses, such as random access queries, frequent incremental updates, and low latency stream processing. The document recommends using a database instead of Spark for random access, updates, and serving live queries. It suggests using message queues instead of files for low latency stream processing. The goal is to help users understand how to properly leverage Spark for big data workloads.
This document discusses data science and the role of data scientists. It defines data science as using theories and principles to perform data-related tasks like collection, cleaning, integration, modeling, and visualization. It distinguishes data science from business intelligence, statistics, database management, and machine learning. Common skills for data scientists include statistics, data munging (formatting data), and visualization. Data scientists perform tasks like preparing models, running models, and communicating results.
Top Big data Analytics tools: Emerging trends and Best practicesSpringPeople
This document discusses top big data analytics tools and emerging trends in big data analytics. It defines big data analytics as examining large data sets to find patterns and business insights. The document then covers several open source and commercial big data analytics tools, including Jaspersoft and Talend for reporting, Skytree for machine learning, Tableau for visualization, and Pentaho and Splunk for reporting. It emphasizes that tool selection is just one part of a big data project and that evaluating business value is also important.
This document provides an overview of modern big data analytics tools. It begins with background on the author and a brief history of Hadoop. It then discusses the growth of the Hadoop ecosystem from early projects like HDFS and MapReduce to a large number of Apache projects and commercial tools. It provides examples of companies and organizations using Hadoop. It also outlines concepts like SQL on Hadoop, in-database analytics using MADLib, and the evolution of Hadoop beyond MapReduce with the introduction of YARN. Finally, it discusses new frameworks being built on top of YARN for interactive, streaming, graph and other types of processing.
Data Culture Series - Keynote - 16th September 2014Jonathan Woodward
Big data. Small data. All data. You have access to an ever-expanding volume of data inside the walls of your business and out across the web. The potential in data is endless – from predicting election results to preventing the spread of epidemics. But how can you use it to your advantage to help move your business forward?
Drive a Data Culture within your organisation
Data Scientist vs Data Analyst vs Data Engineer - Role & Responsibility, Skil...Simplilearn
In this presentation, we will decode the basic differences between data scientist, data analyst and data engineer, based on the roles and responsibilities, skill sets required, salary and the companies hiring them. Although all these three professions belong to the Data Science industry and deal with data, there are some differences that separate them. Every person who is aspiring to be a data professional needs to understand these three career options to select the right one for themselves. Now, let us get started and demystify the difference between these three professions.
We will distinguish these three professions using the parameters mentioned below:
1. Job description
2. Skillset
3. Salary
4. Roles and responsibilities
5. Companies hiring
This Master’s Program provides training in the skills required to become a certified data scientist. You’ll learn the most in-demand technologies such as Data Science on R, SAS, Python, Big Data on Hadoop and implement concepts such as data exploration, regression models, hypothesis testing, Hadoop, and Spark.
Why be a Data Scientist?
Data scientist is the pinnacle rank in an analytics organization. Glassdoor has ranked data scientist first in the 25 Best Jobs for 2016, and good data scientists are scarce and in great demand. As a data scientist you will be required to understand the business problem, design the analysis, collect and format the required data, apply algorithms or techniques using the correct tools, and finally make recommendations backed by data.
Simplilearn's Data Scientist Master’s Program will help you master skills and tools like Statistics, Hypothesis testing, Clustering, Decision trees, Linear and Logistic regression, R Studio, Data Visualization, Regression models, Hadoop, Spark, PROC SQL, SAS Macros, Statistical procedures, tools and analytics, and many more. The courseware also covers a capstone project which encompasses all the key aspects from data extraction, cleaning, visualisation to model building and tuning. These skills will help you prepare for the role of a Data Scientist.
Who should take this course?
The data science role requires the perfect amalgam of experience, data science knowledge, and using the correct tools and technologies. It is a good career choice for both new and experienced professionals. Aspiring professionals of any educational background with an analytical frame of mind are most suited to pursue the Data Scientist Master’s Program, including:
IT professionals
Analytics Managers
Business Analysts
Banking and Finance professionals
Marketing Managers
Supply Chain Network Managers
Those new to the data analytics domain
Students in UG/ PG Analytics Programs
Learn more at https://www.simplilearn.com/big-data-and-analytics/senior-data-scientist-masters-program-training
Extract business value by analyzing large volumes of multi-structured data from various sources such as databases, websites, blogs, social media, smart sensors...
This document provides an overview of big data and big data analytics. It defines big data as large volumes of diverse data that require advanced techniques and technologies to capture, manage, and process within a tolerable time frame. The document outlines the characteristics of big data, including volume, velocity, and variety. It also discusses challenges of big data, examples of big data applications, and different types of analytics including descriptive, predictive, and prescriptive. Recommendation systems are introduced as a type of predictive analytics.
Workshop with Joe Caserta, President of Caserta Concepts, at Data Summit 2015 in NYC.
Data science, the ability to sift through massive amounts of data to discover hidden patterns and predict future trends and actions, may be considered the "sexiest" job of the 21st century, but it requires an understanding of many elements of data analytics. This workshop introduced basic concepts, such as SQL and NoSQL, MapReduce, Hadoop, data mining, machine learning, and data visualization.
For notes and exercises from this workshop, click here: https://github.com/Caserta-Concepts/ds-workshop.
For more information, visit our website at www.casertaconcepts.com
Big Data Presentation - Data Center Dynamics Sydney 2014 - Dez BlanchfieldDez Blanchfield
The document discusses the rise of big data and its impact on data centers. It defines what big data is and what it is not, providing examples of big data sources and uses. It also explores how the concept of a data center is evolving, as they must adapt to support new big data workloads. Traditional data center designs are no longer sufficient and distributed, modular, and software-defined approaches are needed to efficiently manage large and growing volumes of data.
What is Big Data? What is Data Science? What are the benefits? How will they evolve in my organisation?
Built around the premise that the investment in big data is far less than the cost of not having it, this presentation made at a tech media industry event, this presentation will unveil and explore the nuances of Big Data and Data Science and their synergy forming Big Data Science. It highlights the benefits of investing in it and defines a path to their evolution within most organisations.
This document provides an introduction and overview of a summer school course on business analytics and data science. It begins by introducing the instructor and their qualifications. It then outlines the course schedule and topics to be covered, including introductions to data science, analytics, modeling, Google Analytics, and more. Expectations and support resources are also mentioned. Key concepts from various topics are then defined at a high level, such as the data-information-knowledge hierarchy, data mining, CRISP-DM, machine learning techniques like decision trees and association analysis, and types of models like regression and clustering.
Data Visualisation with Hadoop Mashups, Hive, Power BI and Excel 2013Jen Stirrup
The document discusses visualizing big data with tools like Hadoop, Hive, and Excel 2013. It provides an overview of big data technologies and data visualization with Office 365 and Power BI. It describes what Hive is and how it works, including how Hive solves the problem of analyzing large amounts of data by providing a SQL-like language (HiveQL) to query data stored in Hadoop and translating queries to MapReduce jobs. The document demonstrates visualizing big data with Microsoft tools like Power View and Power Map in Excel.
A picture is worth a thousand points of data. The power of data is transformative when the analytics is displayed visually enabling for faster decision making.
Slides from webinar: Provenance and social science data. Presented on 15 March 2017. Presenter was Prof George Alter, Research Professor, ICPSR, and visiting Professor, ANU
FULL webinar recording: https://youtu.be/elPcKqWoOPg
3. Prof George Alter, (Research Professor, ICPSR & Visiting Prof, ANU)
The C2Metadata Project is producing new tools that will work with common statistical packages (eg R and SPSS) to automate the capture of metadata describing variable transformations. Software-independent data transformation descriptions will be added to metadata in two internationally accepted standards: DDI and Ecological Markup Language (EML). These tools will create efficiencies and reduce the costs of data collection, preparation, and re-use. Of special interest to social sciences with its strong metadata standards and heavy reliance on statistical analysis software.
This document discusses visual analytics and big data visualization. It defines big data and explains the need for big data analytics to uncover patterns. Data visualization helps make sense of large datasets and facilitates predictive analysis. Different visualization techniques are described, including charts, graphs, and diagrams suited to simple and big data. Visualization acts as an interface between data storage and users. Characteristics of good visualization and tools for big data visualization are also outlined.
Estamos presenciando inovações tecnológicas que possibilitam utilizar ciência dos dados sem a necessidade de antecipar grandes investimentos. Este contexto facilita a adoção de práticas e valores ágeis que encorajam a antecipação de insights e aprendizado contínuo. Nesta palestra, iremos abordar temas como times multi-funcionais, práticas ágeis de engenharia de software e desenvolvimento iterativo, incremental e colaborativo no contexto de produtos e soluções de ciência dos dados.
Data visualization is often used as the first step while performing a variety of analytical tasks. With the advent of large, high-dimensional datasets and strong interest in data science, there is a need for tools that can support rapid visual analysis. In this paper we describe our vision for a new class of visualization recommendation systems that can automatically identify and interactively recommend visualizations relevant to an analytical task.
This document discusses a presentation on preparing for the ICD-10 transition through documentation, education, and continued training. The presentation will cover ICD-10 compliance issues, coding guidelines, identifying current problem areas, assessing documentation quality, and educating staff on ICD-10 requirements. It will provide guidance on fracture coding, documentation needs, identifying initial vs. subsequent conditions, and using the ICD-10 manual. The goal is to help facilities and their staff successfully address the ICD-10 transition through documentation requirements and an ongoing education process.
Annette Taylor completed the Coursera course "Practical Machine Learning" offered by Johns Hopkins University with distinction. The course taught students the components of machine learning algorithms and how to apply basic machine learning tools to build and evaluate predictors using real data. The certificate was signed by Jeffrey Leek, Roger Peng, and Brian Caffo of Johns Hopkins Bloomberg School of Public Health.
This document discusses data science and the role of data scientists. It defines data science as using theories and principles to perform data-related tasks like collection, cleaning, integration, modeling, and visualization. It distinguishes data science from business intelligence, statistics, database management, and machine learning. Common skills for data scientists include statistics, data munging (formatting data), and visualization. Data scientists perform tasks like preparing models, running models, and communicating results.
Top Big data Analytics tools: Emerging trends and Best practicesSpringPeople
This document discusses top big data analytics tools and emerging trends in big data analytics. It defines big data analytics as examining large data sets to find patterns and business insights. The document then covers several open source and commercial big data analytics tools, including Jaspersoft and Talend for reporting, Skytree for machine learning, Tableau for visualization, and Pentaho and Splunk for reporting. It emphasizes that tool selection is just one part of a big data project and that evaluating business value is also important.
This document provides an overview of modern big data analytics tools. It begins with background on the author and a brief history of Hadoop. It then discusses the growth of the Hadoop ecosystem from early projects like HDFS and MapReduce to a large number of Apache projects and commercial tools. It provides examples of companies and organizations using Hadoop. It also outlines concepts like SQL on Hadoop, in-database analytics using MADLib, and the evolution of Hadoop beyond MapReduce with the introduction of YARN. Finally, it discusses new frameworks being built on top of YARN for interactive, streaming, graph and other types of processing.
Data Culture Series - Keynote - 16th September 2014Jonathan Woodward
Big data. Small data. All data. You have access to an ever-expanding volume of data inside the walls of your business and out across the web. The potential in data is endless – from predicting election results to preventing the spread of epidemics. But how can you use it to your advantage to help move your business forward?
Drive a Data Culture within your organisation
Data Scientist vs Data Analyst vs Data Engineer - Role & Responsibility, Skil...Simplilearn
In this presentation, we will decode the basic differences between data scientist, data analyst and data engineer, based on the roles and responsibilities, skill sets required, salary and the companies hiring them. Although all these three professions belong to the Data Science industry and deal with data, there are some differences that separate them. Every person who is aspiring to be a data professional needs to understand these three career options to select the right one for themselves. Now, let us get started and demystify the difference between these three professions.
We will distinguish these three professions using the parameters mentioned below:
1. Job description
2. Skillset
3. Salary
4. Roles and responsibilities
5. Companies hiring
This Master’s Program provides training in the skills required to become a certified data scientist. You’ll learn the most in-demand technologies such as Data Science on R, SAS, Python, Big Data on Hadoop and implement concepts such as data exploration, regression models, hypothesis testing, Hadoop, and Spark.
Why be a Data Scientist?
Data scientist is the pinnacle rank in an analytics organization. Glassdoor has ranked data scientist first in the 25 Best Jobs for 2016, and good data scientists are scarce and in great demand. As a data scientist you will be required to understand the business problem, design the analysis, collect and format the required data, apply algorithms or techniques using the correct tools, and finally make recommendations backed by data.
Simplilearn's Data Scientist Master’s Program will help you master skills and tools like Statistics, Hypothesis testing, Clustering, Decision trees, Linear and Logistic regression, R Studio, Data Visualization, Regression models, Hadoop, Spark, PROC SQL, SAS Macros, Statistical procedures, tools and analytics, and many more. The courseware also covers a capstone project which encompasses all the key aspects from data extraction, cleaning, visualisation to model building and tuning. These skills will help you prepare for the role of a Data Scientist.
Who should take this course?
The data science role requires the perfect amalgam of experience, data science knowledge, and using the correct tools and technologies. It is a good career choice for both new and experienced professionals. Aspiring professionals of any educational background with an analytical frame of mind are most suited to pursue the Data Scientist Master’s Program, including:
IT professionals
Analytics Managers
Business Analysts
Banking and Finance professionals
Marketing Managers
Supply Chain Network Managers
Those new to the data analytics domain
Students in UG/ PG Analytics Programs
Learn more at https://www.simplilearn.com/big-data-and-analytics/senior-data-scientist-masters-program-training
Extract business value by analyzing large volumes of multi-structured data from various sources such as databases, websites, blogs, social media, smart sensors...
This document provides an overview of big data and big data analytics. It defines big data as large volumes of diverse data that require advanced techniques and technologies to capture, manage, and process within a tolerable time frame. The document outlines the characteristics of big data, including volume, velocity, and variety. It also discusses challenges of big data, examples of big data applications, and different types of analytics including descriptive, predictive, and prescriptive. Recommendation systems are introduced as a type of predictive analytics.
Workshop with Joe Caserta, President of Caserta Concepts, at Data Summit 2015 in NYC.
Data science, the ability to sift through massive amounts of data to discover hidden patterns and predict future trends and actions, may be considered the "sexiest" job of the 21st century, but it requires an understanding of many elements of data analytics. This workshop introduced basic concepts, such as SQL and NoSQL, MapReduce, Hadoop, data mining, machine learning, and data visualization.
For notes and exercises from this workshop, click here: https://github.com/Caserta-Concepts/ds-workshop.
For more information, visit our website at www.casertaconcepts.com
Big Data Presentation - Data Center Dynamics Sydney 2014 - Dez BlanchfieldDez Blanchfield
The document discusses the rise of big data and its impact on data centers. It defines what big data is and what it is not, providing examples of big data sources and uses. It also explores how the concept of a data center is evolving, as they must adapt to support new big data workloads. Traditional data center designs are no longer sufficient and distributed, modular, and software-defined approaches are needed to efficiently manage large and growing volumes of data.
What is Big Data? What is Data Science? What are the benefits? How will they evolve in my organisation?
Built around the premise that the investment in big data is far less than the cost of not having it, this presentation made at a tech media industry event, this presentation will unveil and explore the nuances of Big Data and Data Science and their synergy forming Big Data Science. It highlights the benefits of investing in it and defines a path to their evolution within most organisations.
This document provides an introduction and overview of a summer school course on business analytics and data science. It begins by introducing the instructor and their qualifications. It then outlines the course schedule and topics to be covered, including introductions to data science, analytics, modeling, Google Analytics, and more. Expectations and support resources are also mentioned. Key concepts from various topics are then defined at a high level, such as the data-information-knowledge hierarchy, data mining, CRISP-DM, machine learning techniques like decision trees and association analysis, and types of models like regression and clustering.
Data Visualisation with Hadoop Mashups, Hive, Power BI and Excel 2013Jen Stirrup
The document discusses visualizing big data with tools like Hadoop, Hive, and Excel 2013. It provides an overview of big data technologies and data visualization with Office 365 and Power BI. It describes what Hive is and how it works, including how Hive solves the problem of analyzing large amounts of data by providing a SQL-like language (HiveQL) to query data stored in Hadoop and translating queries to MapReduce jobs. The document demonstrates visualizing big data with Microsoft tools like Power View and Power Map in Excel.
A picture is worth a thousand points of data. The power of data is transformative when the analytics is displayed visually enabling for faster decision making.
Slides from webinar: Provenance and social science data. Presented on 15 March 2017. Presenter was Prof George Alter, Research Professor, ICPSR, and visiting Professor, ANU
FULL webinar recording: https://youtu.be/elPcKqWoOPg
3. Prof George Alter, (Research Professor, ICPSR & Visiting Prof, ANU)
The C2Metadata Project is producing new tools that will work with common statistical packages (eg R and SPSS) to automate the capture of metadata describing variable transformations. Software-independent data transformation descriptions will be added to metadata in two internationally accepted standards: DDI and Ecological Markup Language (EML). These tools will create efficiencies and reduce the costs of data collection, preparation, and re-use. Of special interest to social sciences with its strong metadata standards and heavy reliance on statistical analysis software.
This document discusses visual analytics and big data visualization. It defines big data and explains the need for big data analytics to uncover patterns. Data visualization helps make sense of large datasets and facilitates predictive analysis. Different visualization techniques are described, including charts, graphs, and diagrams suited to simple and big data. Visualization acts as an interface between data storage and users. Characteristics of good visualization and tools for big data visualization are also outlined.
Estamos presenciando inovações tecnológicas que possibilitam utilizar ciência dos dados sem a necessidade de antecipar grandes investimentos. Este contexto facilita a adoção de práticas e valores ágeis que encorajam a antecipação de insights e aprendizado contínuo. Nesta palestra, iremos abordar temas como times multi-funcionais, práticas ágeis de engenharia de software e desenvolvimento iterativo, incremental e colaborativo no contexto de produtos e soluções de ciência dos dados.
Data visualization is often used as the first step while performing a variety of analytical tasks. With the advent of large, high-dimensional datasets and strong interest in data science, there is a need for tools that can support rapid visual analysis. In this paper we describe our vision for a new class of visualization recommendation systems that can automatically identify and interactively recommend visualizations relevant to an analytical task.
This document discusses a presentation on preparing for the ICD-10 transition through documentation, education, and continued training. The presentation will cover ICD-10 compliance issues, coding guidelines, identifying current problem areas, assessing documentation quality, and educating staff on ICD-10 requirements. It will provide guidance on fracture coding, documentation needs, identifying initial vs. subsequent conditions, and using the ICD-10 manual. The goal is to help facilities and their staff successfully address the ICD-10 transition through documentation requirements and an ongoing education process.
Annette Taylor completed the Coursera course "Practical Machine Learning" offered by Johns Hopkins University with distinction. The course taught students the components of machine learning algorithms and how to apply basic machine learning tools to build and evaluate predictors using real data. The certificate was signed by Jeffrey Leek, Roger Peng, and Brian Caffo of Johns Hopkins Bloomberg School of Public Health.
Fraser Island is the largest sand island in the world located off the coast of Queensland, Australia. It is made up of sand that has accumulated over 750,000 years on volcanic bedrock. The island has diverse ecosystems like rainforests, swamps, and coastal dunes. It is home to many plant and animal species. Fraser Island was inhabited by Aboriginal Australians for thousands of years before European settlement in the 1840s disrupted their way of life. The island is now protected as the Great Sandy National Park.
Este documento describe un estudio de tres años sobre las plagas de los espacios verdes urbanos en la ciudad de Lleida, España. El objetivo principal fue determinar las plagas clave, su abundancia relativa y fenología, así como la presencia de enemigos naturales. Los resultados mostraron que los pulgones fueron la plaga más abundante, representando aproximadamente la mitad de las asociaciones cada año. Otras plagas importantes fueron las cochinillas. El estudio proporcionó información básica sobre las plagas que es necesaria para el desarrol
A escola E.M.E.F. 1o de Maio realizou um sarau literário em 2016 para apresentar trabalhos literários dos alunos. O evento ocorreu na cidade de Farroupilha no Rio Grande do Sul e contou com a participação de estudantes.
This document is a portfolio for David Schneider that summarizes his experience in B2B marketing from 2005 to 2007. It outlines his work creating print ads, e-newsletters, packaging, and literature to promote brands to distributors and retailers. It also details how he increased brand awareness through press releases and by networking at trade shows.
People form impressions, or vague ideas, about other people through the process of person perception generally influenced by Physical Appearance or Social or Cognitive schemas
Informacion del programa de mercadeo universidad de los llanosmercadeounillanos
El documento presenta la misión, visión, perfiles de desempeño y profesional, propósitos, objeto de estudio y líneas de profundización del programa de Mercadeo de la Universidad de los Llanos. Su misión es formar ciudadanos éticos y responsables socialmente mediante un currículo flexible e integral que les permita participar en procesos de transformación económica. Su visión es ser reconocido en 2020 por formar profesionales con calidad e investigación que impacten el desarrollo social sostenible.
Este documento resume la profesión de enfermería. Las enfermeras atienden a pacientes, educan a familias, brindan apoyo emocional y realizan procedimientos médicos. Trabajan principalmente en hospitales, clínicas y hogares de ancianos. Se requiere un título universitario en enfermería. Las ventajas incluyen un buen salario, ascensos y la satisfacción de ayudar a otros, mientras que las desventajas son falta de tiempo libre, cansancio y discusiones con familiares.
Cervical cancer is the second most common cancer in women globally. It is caused by persistent infection with human papillomavirus (HPV), with types 16 and 18 causing over 70% of cases. Screening through cervical smears has significantly reduced cervical cancer rates in developed nations. Treatment depends on the stage of cancer, ranging from conization for early stage IA1 to chemoradiation for later stages. Prognosis is best for early stage disease, with 5-year survival rates over 90% for stage IA tumors.
Este documento apresenta os produtos da linha de perfumaria e cosméticos Azenka Cosmetics, incluindo perfumes, desodorantes, hidratantes e produtos para cabelo. A empresa começou suas operações em janeiro de 2015 e já obteve sucesso, oferecendo uma oportunidade de negócio para distribuidores.
La relación que existe entre el mundo odontologico y el consumo de alcohol. En el detalle, hemos dado la prioridad al manejo clinico de estas circunstancias especificas.
Comunicación presentada en el II Simposium Grafica. Encuentro académico de investigación en Diseño Gráfico. Universidad Autónoma de Barcelona, 9 de septiembre de 2016.
La consolidación de los medios informativos digitales supone un reto para la disciplina académica del Diseño Periodístico al que, probablemente, llega tarde. Con un bagaje importante sobre diseño editorial, la orientación en los planes de estudio de Periodismo se ha centrado en el diseño de medios impresos. Mientras tanto, competencias como la investigación con usuarios orientadas a mejorar la arquitectura de información, la navegación o la usabilidad de los interfaces gráficos han quedado fuera, incluso, de disciplinas como la tradicional Tecnología de la Información. Así, parece importante revisar cuáles son las principales competencias que deben adquirir los profesionales del diseño periodístico del Siglo XXI.
Big Data Day LA 2015 - Scalable and High-Performance Analytics with Distribut...Data Con LA
"R is the most popular language in the data-science community with 2+ million users and 6000+ R packages. R’s adoption evolved along with its easy-to-use statistical language, graphics, packages, tools and active community. In this session we will introduce Distributed R, a new open-source technology that solves the scalability and performance limitations of vanilla R. Since R is single-threaded and does not scale to accommodate large datasets, Distributed R addresses many of R’s limitations. Distributed R efficiently shares sparse structured data, leverages multi-cores, and dynamically partitions data to mitigate load imbalance.
In this talk, we will show the promise of this approach by demonstrating how important machine learning and graph algorithms can be expressed in a single framework and are substantially faster under Distributed R. Additionally, we will show how Distributed R complements Vertica, a state-of-the-art columnar analytics database, to deliver a full-cycle, fully integrated, data “prep-analyze-deploy” solution."
R is a programming language and software environment for statistical analysis, graphics, and statistical computing. It is used widely in academic and industry settings. This document provides an introduction to R, including its history and community, how to get started, important data structures, visualization, statistical analysis techniques, and how to work with big data in R. It also discusses challenges of open source R and how Microsoft R products address these challenges through commercial support, scalability, and integration with SQL Server.
Data Science for Fundraising: Build Data-Driven Solutions Using R - Rodger De...Rodger Devine
Although the non-profit industry has advanced using CRMs and donor databases, it has not fully explored the data stored in those databases. Meanwhile, data scientists, in the for-profit industry, using sophisticated tools, have generated data-driven results and effective solutions for several challenges in their organizations. Regardless of your skill level, you can equip yourself and help your organization succeed with these data science techniques using R.
In-Database Analytics Deep Dive with Teradata and RevolutionRevolution Analytics
Teradata and Revolution Analytics worked together to develop in-database analytical capabilities for Teradata Database. Teradata v14.10 provides a foundation for in-database analytics in Teradata. Revolution Analytics has ported its Revolution R Enterprise (RRE) Version 7.1 to use the in-database capabilities of version 14.10. With RRE inside Teradata, users can run fully parallelized algorithms in each node of the Teradata appliance to achieve performance and data scale heretofore unavailable. We'll get past the market-ecture quickly and dive into a “how it really works” presentation, review implications for system configuration and administration, and then take questions from Teradata users who will be charged with deploying and administering Teradata systems as platforms for big data analytics inside the database engine.
microsoft r server for distributed computingBAINIDA
The document introduces Microsoft R Server and Microsoft R Open. It discusses that R is a popular open source programming language and platform for statistics, analytics, and data science. Microsoft R Server allows for distributed computing on big data using R and brings enterprise-grade support and capabilities to the open source R platform. It can perform analytics both in-database using SQL Server and in Hadoop environments without moving data.
Data Science Introduction: Concepts, lifecycle, applications.pptxsumitkumar600840
This document provides an introduction to the subject of data visualization using R programming and Power BI. It discusses key concepts in data science including the data science lifecycle, components of data science like statistics and machine learning, and applications of data science such as image recognition. The document also outlines some advantages and disadvantages of using data science.
FOCUS AREA:
- Identify data requirements and goals.
- IT solution (data) design.
- Focus on data development and configuration (solutions/projects).
- Develop data standards.
- Ensure data integration.
- Ensure correct data testing.
- Maintain and optimize data solutions.
RELATION TO STRATEGY:
- Develop data solutions based on business/IT requirements
Develop data solutions and goals based on operational objectives.
- Link business KPI’s to system KPI’s.
- Ensure correct data reporting in terms of system reports, cockpits, dashboards and scorecards.
This document summarizes a presentation given by Thomas Hütter on using R for data analysis and visualization. The presentation provided an overview of R's history and ecosystem, introduced basic data types and functions, and demonstrated connecting to a SQL Server database to extract and analyze sales data from a Dynamics Nav system. It showed visualizing the results with ggplot2 and creating interactive apps with the Shiny framework. The presentation emphasized that proper data understanding is important for reliable analysis and highlighted resources for learning more about R.
This document provides an overview of Microsoft R and its capabilities for advanced analytics. It discusses how Microsoft R can enable businesses to analyze large volumes of data across multiple environments including locally, on Azure, and with SQL Server and HDInsight. The presentation includes a demonstration of R used with SQL Server, HDInsight, Azure Machine Learning, and Power BI. It highlights how Microsoft R provides a unified platform for data science and analytics that allows users to write code once and deploy models anywhere.
CuRious about R in Power BI? End to end R in Power BI for beginners Jen Stirrup
R is a widely used open-source statistical software environment used by over 2 million data scientists and analysts. It is based on the S programming language and is developed by the R Foundation. R provides a flexible and powerful environment for statistical analysis, modeling, and data visualization. Some key advantages include being free, having an extensive community for support, and allowing for automated replication through scripting. However, it also has some drawbacks like having a steep learning curve and scripts sometimes being difficult to understand.
Training in Analytics, R and Social Media AnalyticsAjay Ohri
This document provides an overview of basics of analysis, analytics, and R. It discusses why analysis is important, key concepts like central tendency, variance, and frequency analysis. It also covers exploratory data analysis, common analytics software, using R for tasks like importing data, data manipulation, visualization and more. Examples and demos are provided for many common R functions and techniques.
The document discusses data mining techniques for predicting currency exchange rates between the US Dollar and Thai Baht. It describes collecting historical data on economic indicators and financial factors from sources like the Bank of Thailand to build a database. Various data mining algorithms like decision trees, naive Bayes, and neural networks are used to analyze the data and identify the most important variables for predicting exchange rates. Graphs show relationships between the Baht exchange rate and factors like gold prices, crude oil prices, stock indexes over 10 years. The goal is to accurately forecast future exchange rates based on the patterns found in the historical data.
The document discusses data warehousing, data mining, and business intelligence applications. It explains that data warehousing organizes and structures data for analysis, and that data mining involves preprocessing, characterization, comparison, classification, and forecasting of data to discover knowledge. The final stage is presenting discovered knowledge to end users through visualization and business intelligence applications.
20150814 Wrangling Data From Raw to Tidy vsIan Feller
This document outlines best practices for processing raw data into tidy datasets. It discusses preparing by validating variables with a codebook, organizing by planning steps and labeling variables, quality control through reproducible code, and communication with comments, codebooks and providing raw and tidy datasets. The presentation demonstrates these practices using examples from agriculture and education data, showing how to reshape data, generate variables, and comment code for clarity.
MSBI online training offered by Quontra Solutions with special features having Extensive Training will be in both MSBI Online Training and Placement. We help you in resume preparation and conducting Mock Interviews.
Emphasis is given on important topics that were required and mostly used in real time projects. Quontra Solutions is an Online Training Leader when it comes to high-end effective and efficient IT Training. We have always been and still are focusing on the key aspect which is providing utmost effective and competent training to both students and professionals who are eager to enrich their technical skills.
Data Science, Statistical Analysis and R... Learn what those mean, how they can help you find answers to your questions and complement the existing toolsets and processes you are currently using to make sense of data. We will explore R and the RStudio development environment, installing and using R packages, basic and essential data structures and data types, plotting graphics, manipulating data frames and how to connect R and SQL Server.
This document provides an overview of exploratory data analysis techniques. It discusses what data is, common sources of data, and different data types and formats. Key steps in exploratory data analysis are introduced, including descriptive statistics, visualizations, and handling messy data. Common measures used to describe central tendency and spread of data are defined. The importance of visualization for exploring relationships and patterns in data is emphasized. Examples of different visualizations are provided.
The document provides an introduction to the R programming language. It discusses that R is an open-source programming language for statistical analysis and graphics. It can run on Windows, Unix and MacOS. The document then covers downloading and installing R and R Studio, the R workspace, basics of R syntax like naming conventions and assignments, working with data in R including importing, exporting and creating calculated fields, using R packages and functions, and resources for R help and tutorials.
DataStax | Data Science with DataStax Enterprise (Brian Hess) | Cassandra Sum...DataStax
Leveraging your operational data for advanced and predictive analytics enables deeper insights and greater value for cloud applications. DSE Analytics is a complete platform for Operational Analytics, including data ingestion, stream processing, batch analysis, and machine learning.
In this talk we will provide an overview of DSE Analytics as it applies to data science tools and techniques, and demonstrate these via real world use cases and examples.
Brian Hess
Rob Murphy
Rocco Varela
About the Speakers
Brian Hess Senior Product Manager, Analytics, DataStax
Brian has been in the analytics space for over 15 years ranging from government to data mining applied research to analytics in enterprise data warehousing and NoSQL engines, in roles ranging from Cryptologic Mathematician to Director of Advanced Analytics to Senior Product Manager. In all these roles he has pushed data analytics and processing to massive scales in order to solve problems that were previously unsolvable.
This document provides an overview of exploratory data analysis (EDA) for machine learning applications. It discusses identifying data sources, collecting data, and the machine learning process. The main part of EDA involves cleaning, preprocessing, and visualizing data to gain insights through descriptive statistics and data visualizations like histograms, scatter plots, and boxplots. This allows discovering patterns, errors, outliers and missing values to understand the dataset before building models.
The Metaverse and AI: how can decision-makers harness the Metaverse for their...Jen Stirrup
The Metaverse is popularized in science fiction, and now it is becoming closer to being a part of our daily lives through the use of social media and shopping companies. How can businesses survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and how does the Metaverse fit into business strategy when futurist ideas are developing into reality at accelerated rates? How do we do this when our data isn't up to scratch? How can we move towards success with our data so we are set up for the Metaverse when it arrives?
How can you help your company evolve, adapt, and succeed using Artificial Intelligence and the Metaverse to stay ahead of the competition? What are the potential issues, complications, and benefits that these technologies could bring to us and our organizations? In this session, Jen Stirrup will explain how to start thinking about these technologies as an organisation.
AI Applications in Healthcare and Medicine.pdfJen Stirrup
This session was delivered for the Global Business Roundtable. The topic: AI applications in Healthcare and Medicine. In this session, Jennifer Stirrup takes people through a general process of adopting AI in their organisations.
BUILDING A STRONG FOUNDATION FOR SUCCESS WITH BI AND DIGITAL TRANSFORMATIONJen Stirrup
The objective of Digital Transformation is improve the quality and resilience of digital services to serve customers better, and data is a cruel part of fulfilling that ambition. As the organisation moves forward in pursuit of its strategic ambitions, it will need to remain focused on the stabilisation and improvement of existing technology and data foundations. To succeed, organisations need continuously strive to improve data, systems and processes for people using digital solutions; it is not simply digitising paper processes. The challenge of digital transformation is to work with people, but how can you build systems that serve them well to achieve and deliver more in a customer-focused way? Innovators will relish the opportunity to adopt new technology, but laggars are often waiting for proof that this will help them deliver better services or products. The challenge is that the adoption of digital solutions varies significantly from one person to the next, one team to the next and one organisation to the next. In this keynote, there will be a discussion of the industry landscape followed by takeaways that will help digital transformation in your organization.
1. Do more than get the basics right
2. Build confidence in changes through better use of data
3. How to oversee delivery while considering strategy
Artificial Intelligence Ethics keynote: With Great Power, comes Great Respons...Jen Stirrup
Artificial Intelligence has been receiving some bad press recently, with respect to its ethical consequences in terms of changes to working conditions, deepfake technology and even job losses. Organizations are concerned about bias in their data, perpetuating stereotypes and neglecting responsibility. How can AI systems treat all people fairly? What about concerns of safety and reliability?
In this keynote, we will explore the toolkits available in Azure to help businesses to navigate the complex ethics environment. Join this session to understand what Microsoft can offer in terms of supporting organisations to consider ethics as an integral part of their AI solutions.
1 Introduction to Microsoft data platform analytics for releaseJen Stirrup
Part 1 of a conference workshop. This forms the morning session, which looks at moving from Business Intelligence to Analytics.
Topics Covered: Azure Data Explorer, Azure Data Factory, Azure Synapse Analytics, Event Hubs, HDInsight, Big Data
Comparing Microsoft Big Data Platform TechnologiesJen Stirrup
In this segment, we look at technologies such as HDInsight, Azure Databricks, Azure Data Lake Analytics and Apache Spark. We compare the technologies to help you to decide the best technology for your situation.
Introduction to Analytics with Azure Notebooks and PythonJen Stirrup
Introduction to Analytics with Azure Notebooks and Python for Data Science and Business Intelligence. This is one part of a full day workshop on moving from BI to Analytics
When looking at Sales Analytics, where should you start? What should you measure? This session provides ideas on sales metrics, implemented in Power BI
This document provides guidance on creating an effective digital marketing analytics dashboard using Power BI. It recommends connecting to Google Analytics as a primary data source and including visualizations of key performance indicators (KPIs) like impressions, clicks, and spending over time. The dashboard should allow users to interact with the data by selecting specific time periods to analyze and compare metrics. Color coding and tooltips can also help users understand relationships in the data and drill down into further details.
Diversity and inclusion for the newbies and doersJen Stirrup
This presentation is aimed at people who want to *do* something positive for diversity and inclusion in their workplaces and communities, but don't know where to start to have a quick impact. I've made up a checklist of 7 'E's to help people along. We cover crucial topics such as: • What can we do to tackle unconscious bias in our systems, solutions and interactions with others? • How can we be more inclusive towards others? • How can we encourage and mentor younger generations to get involved in STEM topics and technical roles both as leaders and in the communities of people who surround us? I hope you enjoy this interactive and thought-provoking discussion of diversity and inclusion, aimed at people who want to get started and do something positive and impactful to help others.
Artificial Intelligence from the Business perspectiveJen Stirrup
What is AI from the Business perspective? In this presentation, Jen Stirrup discusses the 8 'C's of Artificial Intelligence from the business leadership perspective.
How to be successful with Artificial Intelligence - from small to successJen Stirrup
Keynote from AI World Congress in October 2019. Artificial Intelligence isn't just for the technies; it is crucial that business-oriented individuals adopt this technology, which can be conceived as the fourth industrial age. Artificial intelligence is becoming closer to being a a part of our daily lives through the use of technologies like virtual assistants such as Alexa, smart homes, and automated customer service. Now, we are running the race not just to win, but to survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and futurist ideas are developing into reality at accelerated rates.
How can you help your your company to evolve, adapt and succeed using Artificial Intelligence to stay at the forefront of the competition, and win the race for AI adoption in your organization? What are the potential issues, complications and benefits that artificial intelligence could bring to us and our organisations? In this session, Jen Stirrup will explain the quick wins to win the Red Queen's Race in Artificial Intelligence.
Artificial Intelligence: Winning the Red Queen’s Race Keynote at ESPC with Je...Jen Stirrup
Artificial Intelligence is popularised in fiction films such as “The Terminator” and “AI: Artificial Intelligence”. Now, artificial intelligence is becoming closer to being a part of our daily lives through the use of technologies like virtual assistants such as Cortana, smart homes, and automated customer service.
Now, we are running the Red Queen’s race not just to win, but to survive in a world where Artificial Intelligence is becoming the present as well as the future of technology, and futurist ideas are developing into reality at accelerated rates.
How can you help your your company to evolve, adapt and succeed using Artificial Intelligence to stay at the forefront of the competition, and win the Red Queen’s Race? What are the potential issues, complications and benefits that artificial intelligence could bring to us and our organisations?
In this keynote, Jen Stirrup explains the quick wins to win the Red Queen’s Race, using demos from Microsoft technologies such as AutoML to help you and your organisation win the Red Queen’s race.
Data Visualization dataviz superpower! Guidelines on using best practice data visualization principles for Power BI, Excel, SSRS, Tableau and other great tools!
R - what do the numbers mean? #RStats This is the presentation for my Demo at Orlando Live60 AILIve. We go through statistics interpretation with examples
Artificial Intelligence and Deep Learning in Azure, CNTK and TensorflowJen Stirrup
Artificial Intelligence and Deep Learning in Azure, using Open Source technologies CNTK and Tensorflow. The tutorial can be found on GitHub here: https://github.com/Microsoft/CNTK/tree/master/Tutorials
and the CNTK video can be found here: https://youtu.be/qgwaP43ZIwA
Blockchain Demystified for Business Intelligence ProfessionalsJen Stirrup
Blockchain is a transformational technology with the potential to extend digital transformation beyond an organization and into the processes it shares with suppliers, customers, and partners.
What is blockchain? What can it do for my organization? How can your organisation manage a blockchain implementation? How does it work in Azure?
Join this session to learn about blockchain and see it in action. We will also discuss the use cases for blockchain, and whether it is here to stay.
Examples of the worst data visualization everJen Stirrup
This document summarizes an event called SQL Saturday Cork where Jen Stirrup gave a presentation on data visualizations. The document includes objectives for the presentation such as discussing inaccurate data sources and the use of dark colors to represent higher values. It also includes examples of Zimbabwean inflation rates from 1980 to 2008 shown in a table and chart to illustrate how data can be visualized.
Maxx nft market place new generation nft marketing placeusersalmanrazdelhi
PREFACE OF MAXXNFT
MaxxNFT: Powering the Future of Digital Ownership
MaxxNFT is a cutting-edge Web3 platform designed to revolutionize how
digital assets are owned, traded, and valued. Positioned at the forefront of the
NFT movement, MaxxNFT views NFTs not just as collectibles, but as the next
generation of internet equity—unique, verifiable digital assets that unlock new
possibilities for creators, investors, and everyday users alike.
Through strategic integrations with OKT Chain and OKX Web3, MaxxNFT
enables seamless cross-chain NFT trading, improved liquidity, and enhanced
user accessibility. These collaborations make it easier than ever to participate
in the NFT ecosystem while expanding the platform’s global reach.
With a focus on innovation, user rewards, and inclusive financial growth,
MaxxNFT offers multiple income streams—from referral bonuses to liquidity
incentives—creating a vibrant community-driven economy. Whether you
'
re
minting your first NFT or building a digital asset portfolio, MaxxNFT empowers
you to participate in the future of decentralized value exchange.
https://maxxnft.xyz/
Microsoft Build 2025 takeaways in one presentationDigitalmara
Microsoft Build 2025 introduced significant updates. Everything revolves around AI. DigitalMara analyzed these announcements:
• AI enhancements for Windows 11
By embedding AI capabilities directly into the OS, Microsoft is lowering the barrier for users to benefit from intelligent automation without requiring third-party tools. It's a practical step toward improving user experience, such as streamlining workflows and enhancing productivity. However, attention should be paid to data privacy, user control, and transparency of AI behavior. The implementation policy should be clear and ethical.
• GitHub Copilot coding agent
The introduction of coding agents is a meaningful step in everyday AI assistance. However, it still brings challenges. Some people compare agents with junior developers. They noted that while the agent can handle certain tasks, it often requires supervision and can introduce new issues. This innovation holds both potential and limitations. Balancing automation with human oversight is crucial to ensure quality and reliability.
• Introduction of Natural Language Web
NLWeb is a significant step toward a more natural and intuitive web experience. It can help users access content more easily and reduce reliance on traditional navigation. The open-source foundation provides developers with the flexibility to implement AI-driven interactions without rebuilding their existing platforms. NLWeb is a promising level of web interaction that complements, rather than replaces, well-designed UI.
• Introduction of Model Context Protocol
MCP provides a standardized method for connecting AI models with diverse tools and data sources. This approach simplifies the development of AI-driven applications, enhancing efficiency and scalability. Its open-source nature encourages broader adoption and collaboration within the developer community. Nevertheless, MCP can face challenges in compatibility across vendors and security in context sharing. Clear guidelines are crucial.
• Windows Subsystem for Linux is open-sourced
It's a positive step toward greater transparency and collaboration in the developer ecosystem. The community can now contribute to its evolution, helping identify issues and expand functionality faster. However, open-source software in a core system also introduces concerns around security, code quality management, and long-term maintenance. Microsoft’s continued involvement will be key to ensuring WSL remains stable and secure.
• Azure AI Foundry platform hosts Grok 3 AI models
Adding new models is a valuable expansion of AI development resources available at Azure. This provides developers with more flexibility in choosing language models that suit a range of application sizes and needs. Hosting on Azure makes access and integration easier when using Microsoft infrastructure.
Measuring Microsoft 365 Copilot and Gen AI SuccessNikki Chapple
Session | Measuring Microsoft 365 Copilot and Gen AI Success with Viva Insights and Purview
Presenter | Nikki Chapple 2 x MVP and Principal Cloud Architect at CloudWay
Event | European Collaboration Conference 2025
Format | In person Germany
Date | 28 May 2025
📊 Measuring Copilot and Gen AI Success with Viva Insights and Purview
Presented by Nikki Chapple – Microsoft 365 MVP & Principal Cloud Architect, CloudWay
How do you measure the success—and manage the risks—of Microsoft 365 Copilot and Generative AI (Gen AI)? In this ECS 2025 session, Microsoft MVP and Principal Cloud Architect Nikki Chapple explores how to go beyond basic usage metrics to gain full-spectrum visibility into AI adoption, business impact, user sentiment, and data security.
🎯 Key Topics Covered:
Microsoft 365 Copilot usage and adoption metrics
Viva Insights Copilot Analytics and Dashboard
Microsoft Purview Data Security Posture Management (DSPM) for AI
Measuring AI readiness, impact, and sentiment
Identifying and mitigating risks from third-party Gen AI tools
Shadow IT, oversharing, and compliance risks
Microsoft 365 Admin Center reports and Copilot Readiness
Power BI-based Copilot Business Impact Report (Preview)
📊 Why AI Measurement Matters: Without meaningful measurement, organizations risk operating in the dark—unable to prove ROI, identify friction points, or detect compliance violations. Nikki presents a unified framework combining quantitative metrics, qualitative insights, and risk monitoring to help organizations:
Prove ROI on AI investments
Drive responsible adoption
Protect sensitive data
Ensure compliance and governance
🔍 Tools and Reports Highlighted:
Microsoft 365 Admin Center: Copilot Overview, Usage, Readiness, Agents, Chat, and Adoption Score
Viva Insights Copilot Dashboard: Readiness, Adoption, Impact, Sentiment
Copilot Business Impact Report: Power BI integration for business outcome mapping
Microsoft Purview DSPM for AI: Discover and govern Copilot and third-party Gen AI usage
🔐 Security and Compliance Insights: Learn how to detect unsanctioned Gen AI tools like ChatGPT, Gemini, and Claude, track oversharing, and apply eDLP and Insider Risk Management (IRM) policies. Understand how to use Microsoft Purview—even without E5 Compliance—to monitor Copilot usage and protect sensitive data.
📈 Who Should Watch: This session is ideal for IT leaders, security professionals, compliance officers, and Microsoft 365 admins looking to:
Maximize the value of Microsoft Copilot
Build a secure, measurable AI strategy
Align AI usage with business goals and compliance requirements
🔗 Read the blog https://nikkichapple.com/measuring-copilot-gen-ai/
UiPath Community Zurich: Release Management and Build PipelinesUiPathCommunity
Ensuring robust, reliable, and repeatable delivery processes is more critical than ever - it's a success factor for your automations and for automation programmes as a whole. In this session, we’ll dive into modern best practices for release management and explore how tools like the UiPathCLI can streamline your CI/CD pipelines. Whether you’re just starting with automation or scaling enterprise-grade deployments, our event promises to deliver helpful insights to you. This topic is relevant for both on-premise and cloud users - as well as for automation developers and software testers alike.
📕 Agenda:
- Best Practices for Release Management
- What it is and why it matters
- UiPath Build Pipelines Deep Dive
- Exploring CI/CD workflows, the UiPathCLI and showcasing scenarios for both on-premise and cloud
- Discussion, Q&A
👨🏫 Speakers
Roman Tobler, CEO@ Routinuum
Johans Brink, CTO@ MvR Digital Workforce
We look forward to bringing best practices and showcasing build pipelines to you - and to having interesting discussions on this important topic!
If you have any questions or inputs prior to the event, don't hesitate to reach out to us.
This event streamed live on May 27, 16:00 pm CET.
Check out all our upcoming UiPath Community sessions at:
👉 https://community.uipath.com/events/
Join UiPath Community Zurich chapter:
👉 https://community.uipath.com/zurich/
6th Power Grid Model Meetup
Join the Power Grid Model community for an exciting day of sharing experiences, learning from each other, planning, and collaborating.
This hybrid in-person/online event will include a full day agenda, with the opportunity to socialize afterwards for in-person attendees.
If you have a hackathon proposal, tell us when you register!
About Power Grid Model
The global energy transition is placing new and unprecedented demands on Distribution System Operators (DSOs). Alongside upgrades to grid capacity, processes such as digitization, capacity optimization, and congestion management are becoming vital for delivering reliable services.
Power Grid Model is an open source project from Linux Foundation Energy and provides a calculation engine that is increasingly essential for DSOs. It offers a standards-based foundation enabling real-time power systems analysis, simulations of electrical power grids, and sophisticated what-if analysis. In addition, it enables in-depth studies and analysis of the electrical power grid’s behavior and performance. This comprehensive model incorporates essential factors such as power generation capacity, electrical losses, voltage levels, power flows, and system stability.
Power Grid Model is currently being applied in a wide variety of use cases, including grid planning, expansion, reliability, and congestion studies. It can also help in analyzing the impact of renewable energy integration, assessing the effects of disturbances or faults, and developing strategies for grid control and optimization.
Neural representations have shown the potential to accelerate ray casting in a conventional ray-tracing-based rendering pipeline. We introduce a novel approach called Locally-Subdivided Neural Intersection Function (LSNIF) that replaces bottom-level BVHs used as traditional geometric representations with a neural network. Our method introduces a sparse hash grid encoding scheme incorporating geometry voxelization, a scene-agnostic training data collection, and a tailored loss function. It enables the network to output not only visibility but also hit-point information and material indices. LSNIF can be trained offline for a single object, allowing us to use LSNIF as a replacement for its corresponding BVH. With these designs, the network can handle hit-point queries from any arbitrary viewpoint, supporting all types of rays in the rendering pipeline. We demonstrate that LSNIF can render a variety of scenes, including real-world scenes designed for other path tracers, while achieving a memory footprint reduction of up to 106.2x compared to a compressed BVH.
https://arxiv.org/abs/2504.21627
Jira Administration Training – Day 1 : IntroductionRavi Teja
This presentation covers the basics of Jira for beginners. Learn how Jira works, its key features, project types, issue types, and user roles. Perfect for anyone new to Jira or preparing for Jira Admin roles.
Introduction and Background:
Study Overview and Methodology: The study analyzes the IT market in Israel, covering over 160 markets and 760 companies/products/services. It includes vendor rankings, IT budgets, and trends from 2025-2029. Vendors participate in detailed briefings and surveys.
Vendor Listings: The presentation lists numerous vendors across various pages, detailing their names and services. These vendors are ranked based on their participation and market presence.
Market Insights and Trends: Key insights include IT market forecasts, economic factors affecting IT budgets, and the impact of AI on enterprise IT. The study highlights the importance of AI integration and the concept of creative destruction.
Agentic AI and Future Predictions: Agentic AI is expected to transform human-agent collaboration, with AI systems understanding context and orchestrating complex processes. Future predictions include AI's role in shopping and enterprise IT.
nnual (33 years) study of the Israeli Enterprise / public IT market. Covering sections on Israeli Economy, IT trends 2026-28, several surveys (AI, CDOs, OCIO, CTO, staffing cyber, operations and infra) plus rankings of 760 vendors on 160 markets (market sizes and trends) and comparison of products according to support and market penetration.
Droidal: AI Agents Revolutionizing HealthcareDroidal LLC
Droidal’s AI Agents are transforming healthcare by bringing intelligence, speed, and efficiency to key areas such as Revenue Cycle Management (RCM), clinical operations, and patient engagement. Built specifically for the needs of U.S. hospitals and clinics, Droidal's solutions are designed to improve outcomes and reduce administrative burden.
Through simple visuals and clear examples, the presentation explains how AI Agents can support medical coding, streamline claims processing, manage denials, ensure compliance, and enhance communication between providers and patients. By integrating seamlessly with existing systems, these agents act as digital coworkers that deliver faster reimbursements, reduce errors, and enable teams to focus more on patient care.
Droidal's AI technology is more than just automation — it's a shift toward intelligent healthcare operations that are scalable, secure, and cost-effective. The presentation also offers insights into future developments in AI-driven healthcare, including how continuous learning and agent autonomy will redefine daily workflows.
Whether you're a healthcare administrator, a tech leader, or a provider looking for smarter solutions, this presentation offers a compelling overview of how Droidal’s AI Agents can help your organization achieve operational excellence and better patient outcomes.
A free demo trial is available for those interested in experiencing Droidal’s AI Agents firsthand. Our team will walk you through a live demo tailored to your specific workflows, helping you understand the immediate value and long-term impact of adopting AI in your healthcare environment.
To request a free trial or learn more:
https://droidal.com/
European Accessibility Act & Integrated Accessibility TestingJulia Undeutsch
Emma Dawson will guide you through two important topics in this session.
Firstly, she will prepare you for the European Accessibility Act (EAA), which comes into effect on 28 June 2025, and show you how development teams can prepare for it.
In the second part of the webinar, Emma Dawson will explore with you various integrated testing methods and tools that will help you improve accessibility during the development cycle, such as Linters, Storybook, Playwright, just to name a few.
Focus: European Accessibility Act, Integrated Testing tools and methods (e.g. Linters, Storybook, Playwright)
Target audience: Everyone, Developers, Testers
Protecting Your Sensitive Data with Microsoft Purview - IRMS 2025Nikki Chapple
Session | Protecting Your Sensitive Data with Microsoft Purview: Practical Information Protection and DLP Strategies
Presenter | Nikki Chapple (MVP| Principal Cloud Architect CloudWay) & Ryan John Murphy (Microsoft)
Event | IRMS Conference 2025
Format | Birmingham UK
Date | 18-20 May 2025
In this closing keynote session from the IRMS Conference 2025, Nikki Chapple and Ryan John Murphy deliver a compelling and practical guide to data protection, compliance, and information governance using Microsoft Purview. As organizations generate over 2 billion pieces of content daily in Microsoft 365, the need for robust data classification, sensitivity labeling, and Data Loss Prevention (DLP) has never been more urgent.
This session addresses the growing challenge of managing unstructured data, with 73% of sensitive content remaining undiscovered and unclassified. Using a mountaineering metaphor, the speakers introduce the “Secure by Default” blueprint—a four-phase maturity model designed to help organizations scale their data security journey with confidence, clarity, and control.
🔐 Key Topics and Microsoft 365 Security Features Covered:
Microsoft Purview Information Protection and DLP
Sensitivity labels, auto-labeling, and adaptive protection
Data discovery, classification, and content labeling
DLP for both labeled and unlabeled content
SharePoint Advanced Management for workspace governance
Microsoft 365 compliance center best practices
Real-world case study: reducing 42 sensitivity labels to 4 parent labels
Empowering users through training, change management, and adoption strategies
🧭 The Secure by Default Path – Microsoft Purview Maturity Model:
Foundational – Apply default sensitivity labels at content creation; train users to manage exceptions; implement DLP for labeled content.
Managed – Focus on crown jewel data; use client-side auto-labeling; apply DLP to unlabeled content; enable adaptive protection.
Optimized – Auto-label historical content; simulate and test policies; use advanced classifiers to identify sensitive data at scale.
Strategic – Conduct operational reviews; identify new labeling scenarios; implement workspace governance using SharePoint Advanced Management.
🎒 Top Takeaways for Information Management Professionals:
Start secure. Stay protected. Expand with purpose.
Simplify your sensitivity label taxonomy for better adoption.
Train your users—they are your first line of defense.
Don’t wait for perfection—start small and iterate fast.
Align your data protection strategy with business goals and regulatory requirements.
💡 Who Should Watch This Presentation?
This session is ideal for compliance officers, IT administrators, records managers, data protection officers (DPOs), security architects, and Microsoft 365 governance leads. Whether you're in the public sector, financial services, healthcare, or education.
🔗 Read the blog: https://nikkichapple.com/irms-conference-2025/
Supercharge Your AI Development with Local LLMsFrancesco Corti
In today's AI development landscape, developers face significant challenges when building applications that leverage powerful large language models (LLMs) through SaaS platforms like ChatGPT, Gemini, and others. While these services offer impressive capabilities, they come with substantial costs that can quickly escalate especially during the development lifecycle. Additionally, the inherent latency of web-based APIs creates frustrating bottlenecks during the critical testing and iteration phases of development, slowing down innovation and frustrating developers.
This talk will introduce the transformative approach of integrating local LLMs directly into their development environments. By bringing these models closer to where the code lives, developers can dramatically accelerate development lifecycles while maintaining complete control over model selection and configuration. This methodology effectively reduces costs to zero by eliminating dependency on pay-per-use SaaS services, while opening new possibilities for comprehensive integration testing, rapid prototyping, and specialized use cases.
Contributing to WordPress With & Without Code.pptxPatrick Lumumba
Contributing to WordPress: Making an Impact on the Test Team—With or Without Coding Skills
WordPress survives on collaboration, and the Test Team plays a very important role in ensuring the CMS is stable, user-friendly, and accessible to everyone.
This talk aims to deconstruct the myth that one has to be a developer to contribute to WordPress. In this session, I will share with the audience how to get involved with the WordPress Team, whether a coder or not.
We’ll explore practical ways to contribute, from testing new features, and patches, to reporting bugs. By the end of this talk, the audience will have the tools and confidence to make a meaningful impact on WordPress—no matter the skill set.
Introducing FME Realize: A New Era of Spatial Computing and ARSafe Software
A new era for the FME Platform has arrived – and it’s taking data into the real world.
Meet FME Realize: marking a new chapter in how organizations connect digital information with the physical environment around them. With the addition of FME Realize, FME has evolved into an All-data, Any-AI Spatial Computing Platform.
FME Realize brings spatial computing, augmented reality (AR), and the full power of FME to mobile teams: making it easy to visualize, interact with, and update data right in the field. From infrastructure management to asset inspections, you can put any data into real-world context, instantly.
Join us to discover how spatial computing, powered by FME, enables digital twins, AI-driven insights, and real-time field interactions: all through an intuitive no-code experience.
In this one-hour webinar, you’ll:
-Explore what FME Realize includes and how it fits into the FME Platform
-Learn how to deliver real-time AR experiences, fast
-See how FME enables live, contextual interactions with enterprise data across systems
-See demos, including ones you can try yourself
-Get tutorials and downloadable resources to help you start right away
Whether you’re exploring spatial computing for the first time or looking to scale AR across your organization, this session will give you the tools and insights to get started with confidence.
Nix(OS) for Python Developers - PyCon 25 (Bologna, Italia)Peter Bittner
How do you onboard new colleagues in 2025? How long does it take? Would you love a standardized setup under version control that everyone can customize for themselves? A stable desktop setup, reinstalled in just minutes. It can be done.
This talk was given in Italian, 29 May 2025, at PyCon 25, Bologna, Italy. All slides are provided in English.
Original slides at https://slides.com/bittner/pycon25-nixos-for-python-developers
As data privacy regulations become more pervasive across the globe and organizations increasingly handle and transfer (including across borders) meaningful volumes of personal and confidential information, the need for robust contracts to be in place is more important than ever.
This webinar will provide a deep dive into privacy contracting, covering essential terms and concepts, negotiation strategies, and key practices for managing data privacy risks.
Whether you're in legal, privacy, security, compliance, GRC, procurement, or otherwise, this session will include actionable insights and practical strategies to help you enhance your agreements, reduce risk, and enable your business to move fast while protecting itself.
This webinar will review key aspects and considerations in privacy contracting, including:
- Data processing addenda, cross-border transfer terms including EU Model Clauses/Standard Contractual Clauses, etc.
- Certain legally-required provisions (as well as how to ensure compliance with those provisions)
- Negotiation tactics and common issues
- Recent lessons from recent regulatory actions and disputes
2. Note
• This presentation was part of a full day workshop on Power BI and R,
held at SQLBits in 2014
• This is a sample, provided to help you see if my one day Business
Intelligence Masterclass is the right course for you.
• http://bit.ly/BusinessIntelligence2016Masterclass
• In that course, you’ll be given updated notes along with a hands-on
session, so why not join me?
2
3. Course Outline
• Module 1: Setting up your data for R with Power Query
• Module 2: Introducing R
• Module 3: The Big Picture: Putting Power BI and R together
• Module 4: Visualising your data with Power View and Excel 2013
• Module 5: Power Map
• Module 6: Wrap up and Q and Q
4. What is R?
4
• R is a powerful environment for statistical computing
• It is an overgrown calculator
• … which lets you save results in variables
x <- 3
y <- 5
z = 4
x + y + z
5. Vectors in R
5
• create a vector (list) of elements, use the "c" operator
v = c("hello","world","welcome","to","the class.")
v = seq(1,100)
v[1]
v[1:10]
• Subscripting in R square brackets operators allow you to extract values:
• insert logical expressions in the square brackets to retrieve subsets of data from a vector or list. For
example:
6. Vectors in R
Microsoft Confidential 6
v = seq(1,100)
logi = v>95
logi
v[logi]
v[v<6]
v[105]=105
v[is.na(v)]
7. Save and Load RData
Data is saved in R as .Rdata files
Imported back again with load
a <- 1:10
save(a, file = "E:/MyData.Rdata")
rm(a)
load("E:/MyData.Rdata")
print(a)
8. Import From CSV Files
• A simple way to load in data is to read in a CSV.
• Read.csv()
• MyDataFrame <- read.csv(“filepath.csv")
• print(MyDataFrame)
9. Import From CSV Files
• Go to Tools in RStudio, and select Import
Dataset.
• Select the file CountryCodes.csv and select the
Import button.
• In RStudio, you will now see the data in the data
pane.
10. Import From CSV Files
The console window will show the following:
> #import dataset
> CountryCodes <- read.csv("C:/Program Files/R/R-
3.1.0/Working Directory/CountryCodes.csv", header=F)
> View(CountryCodes)
Once the data is imported, we can check the
data.
dim(CountryCodes)
head(CountryCodes)
tail(CountryCodes)
11. Import / Export via ODBC
• The Package RODBC provides R with a connection
to ODBC databases
• library(RODBC)
• myodbcConnect <-
odbcConnect(dsn="servername",uid="us
erid",pwd="******")
12. Import / Export via ODBC
• Query <- "SELECT * FROM lib.table WHERE
..."
• # or read query from file
myQuery <-
readChar("E:/MyQueries/myQuery.sql",
nchars=99999)
myData <- sqlQuery(myodbcConnect,
myQuery, errors=TRUE)
odbcCloseAll()
13. Import/Export from Excel Files
• RODBC also works for importing data from Excel
files
• library(RODBC)
• filename <- "E:/Rtmp/dummmyData.xls"
• myxlsFile <- odbcConnectExcel(filename, readOnly =
FALSE)
• sqlSave(myxlsFile, a, rownames = FALSE)
• b <- sqlFetch(myxlsFile, "a")
• odbcCloseAll()
18. Correlation r = 0.96
18
2000 2001 2002 2003 2004 2005 2006 2007 2008 2009
Number of
people
who died
by
becoming
tangled in
their
bedsheets
Deaths (US)
(CDC)
327 456 509 497 596 573 661 741 809 717
Total
revenue
generated
by skiing
facilities
(US)
Dollars in
millions
(US Census)
1,551 1,635 1,801 1,827 1,956 1,989 2,178 2,257 2,476 2,438
19. R and Power BI together
• Pivot Tables are not always enough
• Scaling Data (ScaleR)
• R is very good at static data visualisation
• Upworthy
19
20. Why R?
• most widely used data analysis software - used by 2M + data
scientist, statisticians and analysts
• Most powerful statistical programming language
• flexible, extensible and comprehensive for productivity
• Create beautiful and unique data visualisations - as seen in New
York Times, Twitter and Flowing Data
• Thriving open-source community - leading edge of analytics
research
• Fills the talent gap - new graduates prefer R.
20
21. Growth in Demand
• Rexer Data Mining survey, 2007 - 2013
• R is the highest paid IT skill Dice.com, Jan 2014
• R most used-data science language after SQL -
O'Reilly, Jan 2014
• R is used by 70% of data miners. Rexer, Sept 2013
21
22. Growth in Demand
• R is #15 of all programming languages.
• RedMonk, Jan 2014
• R growing faster than any other data science
language.
• KDNuggets.
• R is in-memory and limited in the size of data that
you can process.
22
23. What are we testing?
• We have one or two samples and a hypothesis,
which may be true or false.
• The NULL hypothesis – nothing happened.
• The Alternative hypothesis – something did happen.
23
24. Strategy
• We set out to prove that something did happen.
• We look at the distribution of the data.
• We choose a test statistic
• We look at the p value
24
25. How small is too small?
• How do we know when the p-value is small?
• P => 0.05 – Null hypothesis is true
• P < 0.05 – alternative hypothesis is true
• it depends
• For high-risk, then perhaps we want 0.01 or even
0.001.
25
26. Confidence Intervals
• Basically, how confident are you that you can
extrapolate from your little data set to the larger
population?
• We can look at the mean
• To do this, we run a t.test
• t.test(vector)
26
27. Confidence Intervals
• Basically, how confident are you that you can
extrapolate from your little data set to the larger
population?
• We can look at the median
• To do this, we run a Wilcox test.
• t.test(vector)
27
28. Calculate the relative frequency
• How much is above, or below the mean?
• Mean(after > before)
• Mean(abs(x-mean)) < 2 *sd(s)
• This gives you the fraction of data that is greater
than two standard deviations from the mean.
28
29. Testing Categorical Variables for
Independence
• Chi squares – are two variables independent? Are
they connected in some way?
• Summarise the data first: Summary(table(initial,
outcome))
• chisq.test
29
30. How Statistics answers your question
• Is our model significant or insignificant? – The F Statistic
• What is the quality of the model? – R2 statistic
• How well do the data points fit the model? – R2 statistic
31. What do the values mean together?
The type of
analysis
Test statistic How can you tell if it is
significant?
What is the assumption you can make?
Regression analysis F Big F, Small p < 0.05 A general relationship between the
predictors and the response
Regression
Analysis
t Big t (> +2.0
or < -2.0), small p < 0.05
X is an important predictor
Difference of
means
t (two tailed) Big t (> +2.0
or < -2.0), small p < 0.05
Significant difference of means
Difference of
means
t (one tailed) Big t (> +2.0
or < -2.0), small p < 0.05
Significant difference of means
31
32. What is Regression?
Using predictors to predict a response
Using independent variables to predict a dependent variable
Example: Credit score is a response, predicted by spend,
income, location and so on.
33. Linear Regression using World Bank data
We can look at predicting using World Bank data
Year <-
GDP <- (wdiData, )
Plot(wdiData,
Cor(year, wdiData)
Fit <- lm(cpi ~ year+quarter)
Fit
34. Examples of Data Mining in R
cpi2011 <- fit$coefficients[[1]] + fit$coefficients[[2]]*2011 +
fit$coefficients[[3]]*(1:4)
attributes(fit)
fit$coefficients
Residuals(fit) – difference between observed and fitted values
Summary(fit)
Plot(fit)
35. What is Data Mining
Machine Learning
Statistics
Software Engineering and Programming with Data
Intuition
Fun!
36. The Why of Data Mining
to discover new knowledge
to improve business outcomes
to deliver better customised services
37. Examples of Data Mining in R
Logistic Regression (glm)
Decision Trees (rpart, wsrpart)
Random Forests (randomForest, wsrf)
Boosted Stumps (ada)
Neural Networks (nnet)
Support Vector Machines (kernlab)
38. Examples of Data Mining in R
• Packages: – fpc – cluster – pvclust – mclust
• Partitioning-based clustering: kmeans, pam, pamk,
clara
• Hierarchical clustering: hclust, pvclust, agnes, Diana
• Model-based clustering: mclust
• Density-based clustering: dbscan
• Plotting cluster solutions: plotcluster, plot.hclust
• Validating cluster solutions: cluster.stats
40. The Data Mining Process
• Load data
• Choose your variables
• Sample the data into test and training sets (usually 70/30 split)
• Explore the distributions of the data
• Test some distributions
• Transform the data if required
• Build clusters with the data
• Build a model
• Evaluate the model
• Log the data process for auditing externally
41. Loading the Data
• Dsname is the name of our dataset
• Get(dsname)
• Dim(ds)
• Names(ds)
48. Random Forest
• library(randomForest) model <- randomForest(form,
ds[train, vars], na.action=na.omit) model
• ##
• ## Call:
• ## randomForest(formula=form, data=ds[train,
vars], ...
• ## Type of random forest: classification
• ## Number of trees: 500
• ## No. of variables tried at each split: 4 ....
50. Linear Regression
• X: predictor variable
• Y: response variable
• Lm( y ~ x, data= dataframe)
51. Multiple Linear Regression
• Lm is used again
• Lm( y ~ x + u + v, data frame)
• It is better to keep the data in one data
frame because it is easier to manage.
52. Getting Regression Statistics
• Save the model to a variable:
• M <- lm(y ~ x + u + v)
• Then use regression statistics to get the values that you need
from m.
54. Getting regression statistics
• The most important one is summary(m). It shows:
• Estimated coefficients
• Critical statistics such as R2 and the F statistic
• The output is hard to read so we will write it out to Excel.
55. Understanding the Regression Summary
• The model summary gives you the information for
the most important regression statistics, such as the
residuals, coefficients and the significance codes.
• The most important one is the F statistic.
• You can check the residuals whether they are a
normal distribution or not. How can you tell this?
56. Understanding the Regression Summary
• The direction of the median is important e.g. a
negative direction will tell you if there is a skew to
the left.
• The quartiles will also help. Ideally Q1 and Q3 should
have the same magnitude. If not, a skew has
developed. This could be inconsistent with the
median result.
• It helps us to identify outliers.
57. Coefficients and R
• The Estimate column contains estimated regression
coefficients, calculated using the least squares
method. This is the most common method.
• How likely is it that the coefficients are zero? This
only shows estimates. This is the purpose of the
column t and p ( > ¦ t¦)
58. Coefficients and R
• The p value is a probability that this finding is
significant. The lower, the better. We can look at the
column signif. codes to help us to identify the most
appropriate level of p value.
59. Coefficients and R
• R2 is the coefficient of determination. How
successful is the model? We look at this value.
Bigger is better. It is the variance of y that is
explained by the regression model. The remaining
variance is not explained by the model. The adjusted
value takes into account the number of variables in
the model.
60. First Impressions
• Plotting the model can help you to investigate it
further.
• Library(car)
• Outlier.test(m)
• M <- lm(y ~ m)
• Plot(m, which=1)
62. The F Statistic
• Is the model significant or insignificant? This is the purpose of
the F statistic.
• Check the F statistic first because if it is not significant, then the
model doesn’t matter.
63. Significance Stars
The stars are shorthand for significance levels,
with the number of asterisks
displayed according to the p-value computed.
*** for high significance and * for low significance. In
this case, *** indicates that it's unlikely that no
relationship exists b/w heights of parents and heights of
their children.
65. How to get Help
Microsoft Confidential 65
example(rnorm)
Rseek.org
66. Resources
• Introductory Statistics with R by Peter Dalgaard. Good for beginners.
• The Art of R Programming
• http://www.r-project.org
• CRAN sites – Comprehensive R Archive Network