Analiza danych
MrExcel's Holy Macro! Books, Liam Bastick, Oscar...
This book is a practical guide for mastering financial modeling in project finance, providing a clear journey from foundational concepts to advanced techniques. It begins by introducing project finance, its significance, and how it differs from other finance structures. Readers will learn key Excel functions, data validation, and layout strategies essential for creating accurate and dynamic models.As the journey progresses, the book emphasizes best practices for building transparent, flexible, and robust models. It covers linked financial statements, cash flow waterfalls, debt structuring, and valuation techniques. A comprehensive case study walks readers through the construction of a full project finance model, separating construction and operational phases while integrating advanced concepts like scenario planning, sensitivity analysis, and ratio metrics.Designed with a logical flow, this book equips readers with practical skills to tackle real-world financial challenges. From Excel tips to project valuation and funding strategies, it provides actionable insights for analysts, finance professionals, and project managers seeking to excel in project finance modeling.
Financial Modelling in Power BI. Master Subtotals, Functions, and Advanced Excel Tricks in Minutes!
MrExcel's Holy Macro! Books, Jonathan Liau, Liam...
This book introduces readers to the fundamentals of financial modeling using Power BI, starting with an overview of the tool and best practices for creating robust, transparent, and flexible models. Early chapters lay the groundwork by explaining financial statement theory and control accounts, essential concepts for any financial analyst. Readers are guided step-by-step through creating parameters and calculating sales, ensuring a solid foundation in Power BI's core functionalities.As the book progresses, readers delve into more advanced topics such as inventory calculations, operating and capital expenditures, and tax computations. Practical examples and hands-on exercises make complex concepts like DAX functions, FIFO inventory modeling, and control account measures accessible to users of all experience levels. Detailed sections on cash flow statements, income statements, and balance sheets tie the lessons together, showing how these elements integrate into a comprehensive financial model.The final chapters explore advanced features like interest and debt modeling, recursion aversion, and equity calculations, culminating in the creation of fully dynamic and optimized models. Readers also learn to design compelling visualizations to present financial insights effectively. By the end of the journey, users will have the tools and confidence to apply their knowledge to real-world scenarios, mastering financial modeling with Power BI.
Financial Modelling using Dynamic Arrays. Let Lambdas Extend Your Range
MrExcel's Holy Macro! Books, Liam Bastick
Dive into the transformative power of Excel's dynamic arrays in financial modelling. Learn to optimize formulas with LET, create reusable LAMBDA functions, and craft sophisticated models. The book provides a comprehensive introduction to Excel’s dynamic arrays, comparing legacy methodologies with modern capabilities while integrating practical tips and best practices.Through real-world examples and step-by-step tutorials, you’ll uncover the full potential of functions like SORT, FILTER, SEQUENCE, and LAMBDA. Discover how dynamic arrays reduce errors, boost efficiency, and enable innovative approaches to financial modelling. The book also highlights advanced features like eta lambdas and helper functions, offering a deep dive into the cutting-edge tools now available in Excel 365.Whether you’re building complex financial models or just looking to refine your techniques, this guide equips you with the knowledge to transform your processes. Excel enthusiasts and professionals alike will appreciate the clarity and depth this book provides, helping you elevate your modelling game to a whole new level.
Frank Kane
Frank Kane’s Taming Big Data with Apache Spark and Python is your companion to learning Apache Spark in a hands-on manner. Frank will start you off by teaching you how to set up Spark on a single system or on a cluster, and you’ll soon move on to analyzing large data sets using Spark RDD, and developing and running effective Spark jobs quickly using Python.Apache Spark has emerged as the next big thing in the Big Data domain – quickly rising from an ascending technology to an established superstar in just a matter of years. Spark allows you to quickly extract actionable insights from large amounts of data, on a real-time basis, making it an essential tool in many modern businesses. Frank has packed this book with over 15 interactive, fun-filled examples relevant to the real world, and he will empower you to understand the Spark ecosystem and implement production-grade real-time Spark projects with ease.
Fundamentals of Analytics Engineering. An introduction to building end-to-end analytics solutions
Dumky De Wilde, Fanny Kassapian, Jovan Gligorevic,...
Written by a team of 7 industry experts, Fundamentals of Analytics Engineering will introduce you to everything from foundational concepts to advanced skills to get started as an analytics engineer.After conquering data ingestion and techniques for data quality and scalability, you’ll learn about techniques such as data cleaning transformation, data modeling, SQL query optimization and reuse, and serving data across different platforms. Armed with this knowledge, you will implement a simple data platform from ingestion to visualization, using tools like Airbyte Cloud, Google BigQuery, dbt, and Tableau. You’ll also get to grips with strategies for data integrity with a focus on data quality and observability, along with collaborative coding practices like version control with Git. You’ll learn about advanced principles like CI/CD, automating workflows, gathering, scoping, and documenting business requirements, as well as data governance.By the end of this book, you’ll be armed with the essential techniques and best practices for developing scalable analytics solutions from end to end.
Kailash Ahirwar
Generative Adversarial Networks (GANs) have the potential to build next-generation models, as they can mimic any distribution of data. Major research and development work is being undertaken in this field since it is one of the rapidly growing areas of machine learning. This book will test unsupervised techniques for training neural networks as you build seven end-to-end projects in the GAN domain.Generative Adversarial Network Projects begins by covering the concepts, tools, and libraries that you will use to build efficient projects. You will also use a variety of datasets for the different projects covered in the book. The level of complexity of the operations required increases with every chapter, helping you get to grips with using GANs. You will cover popular approaches such as 3D-GAN, DCGAN, StackGAN, and CycleGAN, and you’ll gain an understanding of the architecture and functioning of generative models through their practical implementation.By the end of this book, you will be ready to build, train, and optimize your own end-to-end GAN models at work or in your own projects.
Generative Deep Learning with Python. Unleashing the Creative Power of AI by Mastering AI and Python
Cuantum Technologies LLC
Generative Deep Learning with Python opens the door to the fascinating world of AI where machines create. This course begins with an introduction to deep learning, establishing the essential concepts and techniques. You will then delve into generative models, exploring their theoretical foundations and practical applications. As you progress, you will gain a deep understanding of Generative Adversarial Networks (GANs), learning how they function and how to implement them for tasks like face generation.The course's hands-on projects, such as creating GANs for face generation and using Variational Autoencoders (VAEs) for handwritten digit generation, provide practical experience that reinforces your learning. You'll also explore autoregressive models for text generation, allowing you to see the versatility of generative models across different types of data. Advanced topics will prepare you for cutting-edge developments in the field.Throughout your journey, you will gain insights into the future landscape of generative deep learning, equipping you with the skills to innovate and lead in this rapidly evolving field. By the end of the course, you will have a solid foundation in generative deep learning and be ready to apply these techniques to real-world challenges, driving advancements in AI and machine learning.
Geospatial Data Analytics on AWS. Discover how to manage and analyze geospatial data in the cloud
Scott Bateman, Janahan Gnanachandran, Jeff DeMuth
Managing geospatial data and building location-based applications in the cloud can be a daunting task. This comprehensive guide helps you overcome this challenge by presenting the concept of working with geospatial data in the cloud in an easy-to-understand way, along with teaching you how to design and build data lake architecture in AWS for geospatial data.You’ll begin by exploring the use of AWS databases like Redshift and Aurora PostgreSQL for storing and analyzing geospatial data. Next, you’ll leverage services such as DynamoDB and Athena, which offer powerful built-in geospatial functions for indexing and querying geospatial data. The book is filled with practical examples to illustrate the benefits of managing geospatial data in the cloud. As you advance, you’ll discover how to analyze and visualize data using Python and R, and utilize QuickSight to share derived insights. The concluding chapters explore the integration of commonly used platforms like Open Data on AWS, OpenStreetMap, and ArcGIS with AWS to enable you to optimize efficiency and provide a supportive community for continuous learning.By the end of this book, you’ll have the necessary tools and expertise to build and manage your own geospatial data lake on AWS, along with the knowledge needed to tackle geospatial data management challenges and make the most of AWS services.
Abdishakur Hassan, Jayakrishnan Vijayaraghavan
Data scientists, who have access to vast data streams, are a bit myopic when it comes to intrinsic and extrinsic location-based data and are missing out on the intelligence it can provide to their models. This book demonstrates effective techniques for using the power of data science and geospatial intelligence to build effective, intelligent data models that make use of location-based data to give useful predictions and analyses.This book begins with a quick overview of the fundamentals of location-based data and how techniques such as Exploratory Data Analysis can be applied to it. We then delve into spatial operations such as computing distances, areas, extents, centroids, buffer polygons, intersecting geometries, geocoding, and more, which adds additional context to location data. Moving ahead, you will learn how to quickly build and deploy a geo-fencing system using Python. Lastly, you will learn how to leverage geospatial analysis techniques in popular recommendation systems such as collaborative filtering and location-based recommendations, and more.By the end of the book, you will be a rockstar when it comes to performing geospatial analysis with ease.
Tom Hombergs
We would all like to build software architecture that yields adaptable and flexible software with low development costs. But, unreasonable deadlines and shortcuts make it very hard to create such an architecture.Get Your Hands Dirty on Clean Architecture starts with a discussion about the conventional layered architecture style and its disadvantages. It also talks about the advantages of the domain-centric architecture styles of Robert C. Martin's Clean Architecture and Alistair Cockburn's Hexagonal Architecture. Then, the book dives into hands-on chapters that show you how to manifest a hexagonal architecture in actual code. You'll learn in detail about different mapping strategies between the layers of a hexagonal architecture and see how to assemble the architecture elements into an application. The later chapters demonstrate how to enforce architecture boundaries. You'll also learn what shortcuts produce what types of technical debt and how, sometimes, it is a good idea to willingly take on those debts.After reading this book, you'll have all the knowledge you need to create applications using the hexagonal architecture style of web development.
Michael Hsieh
Amazon SageMaker Studio is the first integrated development environment (IDE) for machine learning (ML) and is designed to integrate ML workflows: data preparation, feature engineering, statistical bias detection, automated machine learning (AutoML), training, hosting, ML explainability, monitoring, and MLOps in one environment.In this book, you'll start by exploring the features available in Amazon SageMaker Studio to analyze data, develop ML models, and productionize models to meet your goals. As you progress, you will learn how these features work together to address common challenges when building ML models in production. After that, you'll understand how to effectively scale and operationalize the ML life cycle using SageMaker Studio.By the end of this book, you'll have learned ML best practices regarding Amazon SageMaker Studio, as well as being able to improve productivity in the ML development life cycle and build and deploy models easily for your ML use cases.
Simon Aubury, Ned Letcher, Kris Jenkins
DuckDB is a fast in-process analytical database. Getting Started with DuckDB offers a practical overview of its usage. You'll learn to load, transform, and query various data formats, including CSV, JSON, and Parquet. The book covers DuckDB's optimizations, SQL enhancements, and extensions for specialized applications. Working with examples in SQL, Python, and R, you'll explore analyzing public datasets and discover tools enhancing DuckDB workflows. This guide suits both experienced and new data practitioners, quickly equipping you to apply DuckDB's capabilities in analytical projects. You'll gain proficiency in using DuckDB for diverse tasks, enabling effective integration into your data workflows.
Asjad Athick
The Elastic Stack helps you work with massive volumes of data to power use cases in the search, observability, and security solution areas.This three-part book starts with an introduction to the Elastic Stack with high-level commentary on the solutions the stack can be leveraged for. The second section focuses on each core component, giving you a detailed understanding of the component and the role it plays. You’ll start by working with Elasticsearch to ingest, search, analyze, and store data for your use cases. Next, you’ll look at Logstash, Beats, and Elastic Agent as components that can collect, transform, and load data. Later chapters help you use Kibana as an interface to consume Elastic solutions and interact with data on Elasticsearch. The last section explores the three main use cases offered on top of the Elastic Stack. You’ll start with a full-text search and look at real-world outcomes powered by search capabilities. Furthermore, you’ll learn how the stack can be used to monitor and observe large and complex IT environments. Finally, you’ll understand how to detect, prevent, and respond to security threats across your environment. The book ends by highlighting architecture best practices for successful Elastic Stack deployments.By the end of this book, you’ll be able to implement the Elastic Stack and derive value from it.
Bhanu Birani
Analytics help you to understand the way your users are interacting with your mobile application. Flurry provides you with a strong analytics tool to keep track of your application and to analyze your application data without much hassle. It provides an SDK for almost all platforms to gather data from all platforms of mobile and web applications. The enormous amount of data Flurry handles directly translates into unique, powerful insights into your applications.Getting Started with Flurry Analytics is a practical, hands-on guide to instantly getting started with Flurry analytics using step-by-step exercises. You will take advantage of the real power of Flurry analytics and provides you with an example-based learning approach to help you master it quickly and efficiently.This book starts by discussing how to set up and integrate Flurry analytics and ends by making you aware of how to track your application using Parse. You will also learn about tracking your application data using Flurry analytics to strengthen your strategies and to improve the application monetization process. You will also learn about configuration and implementation of Flurry analytics and how to make the application track all the necessary details to empower management decisions.
James Church
Every business and organization that collects data is capable of tapping into its own data to gain insights how to improve. Haskell is a purely functional and lazy programming language, well-suited to handling large data analysis problems. This book will take you through the more difficult problems of data analysis in a hands-on manner.This book will help you get up-to-speed with the basics of data analysis and approaches in the Haskell language. You'll learn about statistical computing, file formats (CSV and SQLite3), descriptive statistics, charts, and progress to more advanced concepts such as understanding the importance of normal distribution. While mathematics is a big part of data analysis, we've tried to keep this course simple and approachable so that you can apply what you learn to the real world.By the end of this book, you will have a thorough understanding of data analysis, and the different ways of analyzing data. You will have a mastery of all the tools and techniques in Haskell for effective data analysis.
Alexandre Alves, Lloyd Williams, Robin J. Smith
Events are everywhere, events which can have positive or negative impacts on our lives and important business decisions. These events can impact a company's success, failure, and profitability. Technology now allows people from all walks of life to create Event Driven applications that will immediately and completely respond to the events that affect you and your business. So you are much more responsive to your customers, and competitive threats, and can take advantage of transient time sensitive situations. Getting Started with Oracle Event Processing will let you benefit from the skills and years of experience from the original pioneers who were the driving force behind this immensely flexible, complete, and award winning Event Stream Processing technology. It provides all of the information needed to rapidly deliver and understand Event Driven Architecture (EDA) Applications. These can then be executed on the comprehensive and powerful integral Java Event Server platform which utilizes the hardware and operating system.After an introduction into the benefits and uses of Event Stream Processing, this book uses tutorials and practical examples to teach you how to create valuable and rewarding Event Driven foundational applications. First you will learn how to solve Event Stream Processing problems, followed by the fundamentals of building an Oracle Event processing application in a step by step fashion. Exciting and unique topics are then covered: application construction, the powerful capabilities of the Oracle Event Processing language, CQL, monitoring and managing these applications, and the fascinating domain of real-time Geospatial Movement Analysis. Getting Started with Oracle Event Processing will provide a unique perspective on product creation, evolution and a solid understanding on how to effectively use the product.
Phuong Vo.T.H, Martin Czygan
Data analysis is the process of applying logical and analytical reasoning to study each component of data. Python is a multi-domain, high-level, programming language. It’s often used as a scripting language because of its forgiving syntax and operability with a wide variety of different eco-systems. Python has powerful standard libraries or toolkits such as Pylearn2 and Hebel, which offers a fast, reliable, cross-platform environment for data analysis.With this book, we will get you started with Python data analysis and show you what its advantages are.The book starts by introducing the principles of data analysis and supported libraries, along with NumPy basics for statistic and data processing. Next it provides an overview of the Pandas package and uses its powerful features to solve data processing problems.Moving on, the book takes you through a brief overview of the Matplotlib API and some common plotting functions for DataFrame such as plot. Next, it will teach you to manipulate the time and data structure, and load and store data in a file or database using Python packages. The book will also teach you how to apply powerful packages in Python to process raw data into pure and helpful data using examples.Finally, the book gives you a brief overview of machine learning algorithms, that is, applying data analysis results to make decisions or build helpful products, such as recommendations and predictions using scikit-learn.
Luca Zamboni
Simulink is an engineer's Swiss army knife: instead of spending the day typing out complex formulas, Simulink enables you to both draw and execute them. Block after block, you can develop your ideas without struggling with obscure programming languages and you don't have to wait to debug your algorithm - just launch a simulation!Getting Started with Simulink will give you comprehensive knowledge of Simulink's capabilities. From the humble constant block to the S-function block, you will have a clear understanding of what modelling really means, without feeling that something has been left out. By the time you close the book, you'll be able to further extend your modelling skills without any help.We''ll start with a brief introduction, and immediately start placing the first blocks. Little by little, you'll build a car cruise controller model, followed by the mathematical model of a sports car in order to calibrate it. Then you'll learn how to interface your Simulink model with the external world. This book will give you an easy understanding of the tools Simulink offers you, guiding you through a complex exercise split into the three main phases of Simulink development: modelling, testing, and interfacing.
Tyler Richards
Streamlit shortens the development time for the creation of data-focused web applications, allowing data scientists to create web app prototypes using Python in hours instead of days. Getting Started with Streamlit for Data Science takes a hands-on approach to helping you learn the tips and tricks that will have you up and running with Streamlit in no time.You'll start with the fundamentals of Streamlit by creating a basic app and gradually build on the foundation by producing high-quality graphics with data visualization and testing machine learning models. As you advance through the chapters, you’ll walk through practical examples of both personal data projects and work-related data-focused web applications, and get to grips with more challenging topics such as using Streamlit Components, beautifying your apps, and quick deployment of your new apps.By the end of this book, you’ll be able to create dynamic web apps in Streamlit quickly and effortlessly using the power of Python.
Tristan Guillevin
Tableau is one of the leading business intelligence tools used worldwide, in organizations of every scale. In its latest release, Tableau 2018 promises richer and more useful features related to visual analytics, reporting, dashboarding, and a host of other data visualization aspects. Getting Started with Tableau 2018.x will get you up and running with these features.The book starts with all the new functionalities of the different Tableau 2018 versions, along with concrete examples of how to use them. However, if you're new to Tableau, don't worry! The rest of the book will guide you through each major aspect of Tableau with examples. You'll learn how to connect to data, build a data source, visualize your data, build a dashboard, and share it online. In the final chapters, you'll also learn advanced techniques such as creating a cross-database join, data blending, and more.By the end of the book, you will have a firm understanding of how to effectively use Tableau to create quick, cost-effective, and business-efficient business intelligence solutions.
Tristan Guillevin
Tableau is one of the leading data visualization tools and is regularly updated with new functionalities and features. The latest release, Tableau 2019.2, promises new and advanced features related to visual analytics, reporting, dashboarding, and a host of other data visualization aspects. Getting Started with Tableau 2019.2 will get you up to speed with these additional functionalities.The book starts by highlighting the new functionalities of Tableau 2019.2, providing concrete examples of how to use them. However, if you're new to Tableau, you won't have to worry as the book also covers the major aspects of Tableau with relevant examples. You'll learn how to connect to data, build a data source, visualize your data, build a dashboard, and even share data online. In the concluding chapters, you'll delve into advanced techniques such as creating a cross-database join and data blending.By the end of this book, you will be able to use Tableau effectively to create quick, cost-effective, and business-efficient Business Intelligence (BI) solutions.
Donato Teutonico
This book is perfect for R programmers who are interested in learning to use ggplot2 for data visualization, from the basics up to using more advanced applications, such as faceting and grouping. Since this book will not cover the basics of R commands and objects, you should have a basic understanding of the R language.
Go Design Patterns. Best practices in software development and CSP
Mario Castro Contreras
Go is a multi-paradigm programming language that has built-in facilities to create concurrent applications. Design patterns allow developers to efficiently address common problems faced during developing applications. Go Design Patterns will provide readers with a reference point to software design patterns and CSP concurrency design patterns to help them build applications in a more idiomatic, robust, and convenient way in Go. The book starts with a brief introduction to Go programming essentials and quickly moves on to explain the idea behind the creation of design patterns and how they appeared in the 90’s as a common language between developers to solve common tasks in object-oriented programming languages. You will then learn how to apply the 23 Gang of Four (GoF) design patterns in Go and also learn about CSP concurrency patterns, the killer feature in Go that has helped Google develop software to maintain thousands of servers. With all of this the book will enable you to understand and apply design patterns in an idiomatic way that will produce concise, readable, and maintainable software.
Go Web Scraping Quick Start Guide. Implement the power of Go to scrape and crawl data from the web
Vincent Smith
Web scraping is the process of extracting information from the web using various tools that perform scraping and crawling. Go is emerging as the language of choice for scraping using a variety of libraries. This book will quickly explain to you, how to scrape data data from various websites using Go libraries such as Colly and Goquery.The book starts with an introduction to the use cases of building a web scraper and the main features of the Go programming language, along with setting up a Go environment. It then moves on to HTTP requests and responses and talks about how Go handles them. You will also learn about a number of basic web scraping etiquettes.You will be taught how to navigate through a website, using a breadth-first and then a depth-first search, as well as find and follow links. You will get to know about the ways to track history in order to avoid loops and to protect your web scraper using proxies.Finally the book will cover the Go concurrency model, and how to run scrapers in parallel, along with large-scale distributed web scraping.