of critical information infrastructure objects, as well as telecommunication networks used to organize the interaction of critical information infrastructure objects with each other.
Cross-entropy (Кросс-энтропия) – A generalization of Log Loss to multi-class classification problems. Cross-entropy quantifies the difference between two probability distributions. See also perplexity.
Crossover (Also recombination) (Кроссовер) – In genetic algorithms and evolutionary computation, a genetic operator used to combine the genetic information of two parents to generate new offspring. It is one way to stochastically generate new solutions from an existing population, and analogous to the crossover that happens during sexual reproduction in biological organisms. Solutions can also be generated by cloning an existing solution, which is analogous to asexual reproduction. Newly generated solutions are typically mutated before being added to the population [127].
Cross-Validation (k-fold Cross-Validation, Leave-p-out Cross-Validation) (Перекрёстная проверка) – A collection of processes designed to evaluate how the results of a predictive model will generalize to new data sets. k-fold Cross-Validation; Leave-p-out Cross-Validation.
Cryogenic freezing (cryonics, human cryopreservation) is a technology of preserving in a state of deep cooling (using liquid nitrogen) the head or body of a person after his death with the intention to revive them in the future.
Cyber-physical systems (Киберфизические системы) are intelligent networked systems with built-in sensors, processors and drives that are designed to interact with the physical environment and support the operation of computer information systems in real time; cloud computing is an information technology model for providing ubiquitous and convenient access using the information and telecommunications network “Internet” to a common set of configurable computing resources (“cloud”), data storage devices, applications and services that can be promptly provided and relieved from the load with minimal operating costs or almost without the participation of the provider.
Cyber-physical systems (Киберфизические системы) are intelligent networked systems with built-in sensors, processors and drives that are designed to interact with the physical environment and support the operation of computer information systems in real time; cloud computing is an information technology model for providing ubiquitous and convenient access using the information and telecommunications network “Internet” to a common set of configurable computing resources (“cloud”), data storage devices, applications and services that can be promptly provided and relieved from the load with minimal operating costs or almost without the participation of the provider.
“D”
Darkforest (Программа Darkforest) – A computer program, based on deep learning techniques using a convolutional neural network. Its updated version Darkforest2 combines the techniques of its predecessor with Monte Carlo tree search. The MCTS effectively takes tree search methods commonly seen in computer chess programs and randomizes them. With the update, the system is known as Darkforest3.
Dartmouth workshop (Дартмутский семинар) – The Dartmouth Summer Research Project on Artificial Intelligence was the name of a 1956 summer workshop now considered by many (though not all) to be the seminal event for artificial intelligence as a field.
Data (Данные) – Data is a collection of qualitative and quantitative variables. It contains the information that is represented numerically and needs to be analyzed.
Data analysis (Анализ данных) – Obtaining an understanding of data by considering samples, measurement, and visualization. Data analysis can be particularly useful when a dataset is first received, before one builds the first model. It is also crucial in understanding experiments and debugging problems with the system [128].
Data analytics (Аналитика данных)
Data analytics is the science of analyzing raw data to make conclusions about that information. Many of the techniques and processes of data analytics have been automated into mechanical processes and algorithms that work over raw data for human consumption. [129]
Data augmentation (Увеличение данных в анализе данных) – Data augmentation in data analysis are techniques used to increase the amount of data. It helps reduce overfitting when training a machine learning [130].
Data Cleaning (Очистка данных) – Data Cleaning is the process of identifying, correcting, or removing inaccurate or corrupt data records.
Data Curation (Курирование данных) – Data Curation includes the processes related to the organization and management of data which is collected from various sources [131].
Data entry (Ввод данных) The process of converting verbal or written responses to electronic form. [132]
Data fusion (Слияние данных) — The process of integrating multiple data sources to produce more consistent, accurate, and useful information than that provided by any individual data source [].
Data Integration (Интеграция данных) – involves the combination of data residing in different resources and then the supply in a unified view to the users. Data integration is in high demand for both commercial and scientific domains in which they need to merge the data and research results from different repositories [].
Data Lake (Озеро данных) – A type of data repository that stores data in its natural format and relies on various schemata and structure to index the data.
Data markup (Разметка данных) is the stage of processing structured and unstructured data, during which data (including text documents, photo and video images) are assigned identifiers that reflect the type of data (data classification), and (or) data is interpreted to solve a specific problem, in including using machine learning methods (National Strategy for the Development of Artificial Intelligence for the period up to 2030).
Data Mining (Интеллектуальный анализ данных) – is the process of data analysis and information extraction from large amounts of datasets with machine learning, statistical approaches. and many others. [133]
Data parallelism (Параллелизм данных) – A way of scaling training or inference that replicates an entire model onto multiple devices and then passes a subset of the input data to each device. Data parallelism can enable training and inference on very large batch sizes; however, data parallelism requires that the model be small enough to fit on all devices. See also model parallelism.
Data protection (Защита данных) is the process of protecting data and involves the relationship between the collection and dissemination of data and technology, the public perception and expectation of privacy and the political and legal underpinnings surrounding that data. It aims to strike