What does big data in the book 'Big Data Era' mean?
Big data in the book "Big Data Era" refers to "all data or all data", also known as "huge data", which refers to the amount of data involved that is so huge that it cannot be passed through the current mainstream Software tools can capture, manage, process, and organize information within a reasonable time to help enterprises make more positive business decisions.
The operating environment of this tutorial: Windows 7 system, Dell G3 computer.
Big data in the book "Big Data Era" refers to "all data or all data".
Big data (big data), or huge amount of data, refers to the amount of data involved that is so large that it cannot be captured, managed, and managed within a reasonable time by current mainstream software tools. Process and organize information to help companies make more positive business decisions.
In "The Age of Big Data" written by Victor Meyer-Schoenberg and Kenneth Cukier, big data refers to the use of all data without shortcuts such as random analysis (sampling survey). method) 4V characteristics of big data: Volume, Velocity, Variety, and Value.
History of the development of the concept of big data:
The earliest reference to the term "big data" can be traced back to the open source project Nutch of apache org. At the time, big data was used to describe large data sets that needed to be batch processed or analyzed simultaneously to update web search indexes. With the release of Google MapReduce and Google File System (GFS), big data is no longer just used to describe large amounts of data, but also covers the speed of processing data.
As early as 1980, the famous futurist Alvin Toffler enthusiastically praised big data as "the cadenza of the third wave" in his book "The Third Wave". .
However, starting around 2009, “163 big data” became a popular vocabulary in the Internet information technology industry. The U.S. Internet Data Center pointed out that the data on the Internet will grow by 50% every year and double every two years. At present, more than 90% of the data in the world was generated in recent years. In addition, data does not simply refer to the information people publish on the Internet. There are countless digital sensors on industrial equipment, cars, and electricity meters around the world, measuring and transmitting information about position, movement, vibration, temperature, humidity, and even chemistry in the air at any time. Changes in matter also generate massive amounts of data information.
Conceptual structure of big data:
Big data is just a manifestation or characteristic of the development of the Internet to the present stage. There is no need to myth it or maintain awe of it. , with the backdrop of technological innovation represented by cloud computing, these data that were originally difficult to collect and use have begun to be easily utilized. Through continuous innovation in all walks of life, big data will gradually create more for human beings. value.
Secondly, if you want to systematically recognize big data, you must decompose it comprehensively and carefully. I will start from three levels:
The first level is theory, and theory is recognition. It is the only way to know, and it is also the baseline that is widely recognized and disseminated. I will understand the industry's overall description and characterization of big data from the definition of the characteristics of big data; deeply analyze the preciousness of big data from the discussion of the value of big data; gain insight into the development trend of big data; and start from the special and important issue of big data privacy. Examine the long-term game between people and data from a perspective.
The second level is technology. Technology is the means to embody the value of big data and the cornerstone of progress. I will explain the entire process of big data from collection, processing, storage to result formation from the development of cloud computing, distributed processing technology, storage technology and perception technology respectively.
The third level is practice, and practice is the ultimate value manifestation of big data. I will describe the beautiful scene that big data has shown and the blueprint for its upcoming realization from four aspects: Internet big data, government big data, enterprise big data and personal big data.
Characteristics of the big data concept:
Compared with traditional data warehouse applications, big data analysis has the characteristics of large data volume and complex query and analysis. The article "Architecting Big Data: Challenges, Current Situation and Prospects" published in "Journal of Computer Science" lists several important features that a big data analysis platform needs to have, and analyzes the current mainstream implementation platforms - parallel databases, MapReduce and hybrids based on the two. The architecture is analyzed and summarized, and their respective advantages and disadvantages are pointed out. At the same time, the current research status of each direction and the author's efforts in big data analysis are introduced, and future research is prospected.
The four "Vs" or characteristics of big data have four levels: First, the volume of data is huge. From the TB level to the PB level; second, there are many data types. The web logs, videos, pictures, geographical location information, etc. mentioned above. Third, the processing speed is fast and the 1-second rule can quickly obtain high-value information from various types of data. This is also fundamentally different from traditional data mining technology. Fourth, as long as the data is properly utilized and analyzed correctly and accurately, it will bring high value returns. The industry summarizes it into four "V" - Volume, Variety, Velocity, and Value.
To some extent, big data is the cutting-edge technology of data analysis. In short, the ability to quickly obtain valuable information from various types of data is big data technology. Understanding this is critical, and it’s what drives this technology’s potential to reach so many businesses.
Use of big data concept:
Big data can be divided into fields such as big data technology, big data engineering, big data science and big data application. What people are talking about the most now is big data technology and big data applications. Engineering and scientific issues have not yet been taken seriously. Big data engineering refers to the systematic engineering of planning, construction, operation and management of big data; big data science focuses on discovering and verifying the laws of big data and its relationship with natural and social activities during the development and operation of big data networks.
The Internet of Things, cloud computing, mobile Internet, Internet of Vehicles, mobile phones, tablets, PCs, and various sensors spread across every corner of the earth are all data sources or carrying methods.
Some examples include weblogs, RFID, sensor networks, social networks, social data (thanks to the data revolution in society), Internet text and files; Internet search indexes; call detail logging, astronomy, atmospheric science, genomics , biogeochemical, biological, and other complex and/or interdisciplinary scientific research, military reconnaissance, medical records; photographic archives, video archives; and large-scale electronic commerce.
The role of big data
For general enterprises, the role of big data is mainly reflected in two aspects, namely the analysis and use of data and secondary processing development projects. By analyzing the big data of Xijin Information, we can not only dig out hidden data, but also use these hidden messages to improve our customer base through physical sales. As for the secondary development of data, it is often used in network service projects. By summarizing and analyzing this information, we can develop personalized plans that meet customer needs and create a new advertising and marketing method. What you need to understand here is that combining products and services through big data analysis is not an accident. Those who realize this are often leaders in the data era.
To sum up, the application of big data not only marks the progress of the times, but also inspires people to conduct deeper exploration. In addition, for the research on big data, in addition to the above content, it is also necessary to understand the three characteristics of big data, namely large scale, fast operation speed and data diversity. By studying these three aspects, it is not only easier to observe the nature of the data, but also conducive to the effective operation of the software processing platform.
For more related knowledge, please visit the FAQ column!
The above is the detailed content of What does big data in the book 'Big Data Era' mean?. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



Big data structure processing skills: Chunking: Break down the data set and process it in chunks to reduce memory consumption. Generator: Generate data items one by one without loading the entire data set, suitable for unlimited data sets. Streaming: Read files or query results line by line, suitable for large files or remote data. External storage: For very large data sets, store the data in a database or NoSQL.

In the Internet era, big data has become a new resource. With the continuous improvement of big data analysis technology, the demand for big data programming has become more and more urgent. As a widely used programming language, C++’s unique advantages in big data programming have become increasingly prominent. Below I will share my practical experience in C++ big data programming. 1. Choosing the appropriate data structure Choosing the appropriate data structure is an important part of writing efficient big data programs. There are a variety of data structures in C++ that we can use, such as arrays, linked lists, trees, hash tables, etc.

AEC/O (Architecture, Engineering & Construction/Operation) refers to the comprehensive services that provide architectural design, engineering design, construction and operation in the construction industry. In 2024, the AEC/O industry faces changing challenges amid technological advancements. This year is expected to see the integration of advanced technologies, heralding a paradigm shift in design, construction and operations. In response to these changes, industries are redefining work processes, adjusting priorities, and enhancing collaboration to adapt to the needs of a rapidly changing world. The following five major trends in the AEC/O industry will become key themes in 2024, recommending it move towards a more integrated, responsive and sustainable future: integrated supply chain, smart manufacturing

1. Background of the Construction of 58 Portraits Platform First of all, I would like to share with you the background of the construction of the 58 Portrait Platform. 1. The traditional thinking of the traditional profiling platform is no longer enough. Building a user profiling platform relies on data warehouse modeling capabilities to integrate data from multiple business lines to build accurate user portraits; it also requires data mining to understand user behavior, interests and needs, and provide algorithms. side capabilities; finally, it also needs to have data platform capabilities to efficiently store, query and share user profile data and provide profile services. The main difference between a self-built business profiling platform and a middle-office profiling platform is that the self-built profiling platform serves a single business line and can be customized on demand; the mid-office platform serves multiple business lines, has complex modeling, and provides more general capabilities. 2.58 User portraits of the background of Zhongtai portrait construction

In today's big data era, data processing and analysis have become an important support for the development of various industries. As a programming language with high development efficiency and superior performance, Go language has gradually attracted attention in the field of big data. However, compared with other languages such as Java and Python, Go language has relatively insufficient support for big data frameworks, which has caused trouble for some developers. This article will explore the main reasons for the lack of big data framework in Go language, propose corresponding solutions, and illustrate it with specific code examples. 1. Go language

Yizhiwei’s 2023 autumn product launch has concluded successfully! Let us review the highlights of the conference together! 1. Intelligent inclusive openness, allowing digital twins to become productive Ning Haiyuan, co-founder of Kangaroo Cloud and CEO of Yizhiwei, said in his opening speech: At this year’s company’s strategic meeting, we positioned the main direction of product research and development as “intelligent inclusive openness” "Three core capabilities, focusing on the three core keywords of "intelligent inclusive openness", we further proposed the development goal of "making digital twins a productive force". 2. EasyTwin: Explore a new digital twin engine that is easier to use 1. From 0.1 to 1.0, continue to explore the digital twin fusion rendering engine to have better solutions with mature 3D editing mode, convenient interactive blueprints, and massive model assets

As an open source programming language, Go language has gradually received widespread attention and use in recent years. It is favored by programmers for its simplicity, efficiency, and powerful concurrent processing capabilities. In the field of big data processing, the Go language also has strong potential. It can be used to process massive data, optimize performance, and can be well integrated with various big data processing tools and frameworks. In this article, we will introduce some basic concepts and techniques of big data processing in Go language, and show how to use Go language through specific code examples.

In big data processing, using an in-memory database (such as Aerospike) can improve the performance of C++ applications because it stores data in computer memory, eliminating disk I/O bottlenecks and significantly increasing data access speeds. Practical cases show that the query speed of using an in-memory database is several orders of magnitude faster than using a hard disk database.