Learning Centre
Like artificial intelligence, human beings progress through learning. Digital technology, the cloud and the internet in general are vast, fascinating subjects. This is why we dedicate this Glossary Directory to those who wish to discover, learn about or get started with current and future technologies. Develop new skills in areas such as cloud computing, IT security and big data using digital solutions.

Application integration and modernisation
Artificial intelligence (AI) is becoming increasingly linked to the digital technologies we use every day, and has consistently improved their effectiveness. AI assimilates a lot of information with the data it receives. It then analyses this data to deliver the best possible action.
An application programming interface (API) is a software interface that enables multiple applications and services to exchange data. The API can also be used to launch access requests from one application to another, so different services can be connected to each other.
A REST API is a set of architectural rules for implementing web services. This type of API is often used as it increases simplicity, scalability, performance and reliability.
Many businesses are reliant on legacy applications, leaving them unable to leverage cloud computing’s huge potential. By modernising existing software using a cloud-first approach, enterprise can improve performance and scalability while opening up new opportunities for growth, productivity and ways of working.
If an enterprise operations are disrupted by disaster, the right business continuity strategy ensures there are processes in place for the company to continue operating effectively. Crucially, business continuity gives the organisation time and breathing space to address the source of interruption.
Quantum machine learning is a fusion of machine learning and quantum computing, which uses quantum computers to analyse complex data and solve problems at unprecedented speeds, opening up new perspectives in fields such as the pharmaceutical industry, finance, and technology.
Generative AI, a branch of AI focusing on content creation, uses neural networks to generate new output from prompts. It learns from vast data to produce text, images, and more, mimicking human creativity. Businesses use it for marketing, customer support, and analytics, enhancing productivity and decision-making.
The SDLC (Software Development Lifecycle) is a methodology used to develop software applications. It enables developers to build and deploy high-quality software that meets customer expectations, whilst also being delivered on time and within budget.
Extensible Markup Language (XML) uses tags to tell how the text in a data file should be structured, stored and transported. Designed to be readable by both humans and machines, XML is a powerful, highly customisable tagging tool ideally suited for the data-driven era.
CI/CD, or continuous integration and continuous deployment, is a software development practice that automates the processes of developing, testing, and deploying applications. It aims to quickly detect errors, ensure regular and reliable code delivery, and accelerate the development cycle.
SaaS (Software as a Service) is a cloud computing model where users access software hosted online. The publisher manages the software, and the provider, such as OVHcloud, manages the infrastructure.
AI prompt engineering combines art and science, guiding language models to produce the best outcomes. It's akin to choosing words carefully to instruct a brilliant yet literal-minded friend. Prompt engineers apply their AI understanding and creativity to craft queries for generating text, translating languages, and more, ensuring informative and inventive responses.
Service-Oriented Architecture (SOA) is a type of software design that makes services reusable and interoperable using service interfaces. SOA enables businesses to leverage services that can be combined and reconfigured rapidly to meet changing requirements.
Microservices architecture is a software design approach that structures applications as a collection of small, independent services. Each service focuses on a specific function, communicating through APIs. This approach offers increased flexibility, scalability, resilience, and faster development cycles compared to traditional monolithic architectures.
This comprehensive guide demystifies neural networks. Learn about their brain-inspired design, components like neurons and layers, and training methods involving forward and backpropagation. Explore diverse architectures and their applications in image recognition, language processing, forecasting, and more. Discover how to train these powerful tools and their impact across industries.
SAP is a leading provider of enterprise resource planning (ERP) software, offering integrated solutions for finance, logistics, HR, CRM, and more. Its evolution from SAP R/1 to SAP S/4HANA reflects its commitment to innovation and adapting to changing business needs.
Middleware is crucial for software communication and data exchange. It simplifies development, enables interoperability, and enhances scalability. The right web host will provide you with load balancers, managed Kubernetes, and object storage, enabling integration into middleware architecture.
Ubuntu is one of the most popular Linux-based operating systems, known for its user-friendly interface, robust security features, and strong community support. In this guide, we'll dive deep into the meaning of Ubuntu, its uses, and why it has become a favorite among tech enthusiasts, developers, and businesses worldwide.
SAP S/4HANA is an advanced ERP suite designed to enhance business efficiency through real-time insights, simplified processes, and innovative technology. Built on the SAP HANA database, it offers agile, scalable solutions for finance, supply chain, and more.
SAP HANA is a revolutionary in-memory database platform that empowers businesses with real-time data processing and advanced analytics. Its speed, scalability, and integration capabilities make it ideal for driving innovation and digital transformation.
IaaS (Infrastructure as a Service) is a cloud computing solution that offers on-demand computing resources, such as computing power and storage, without any hardware management. It saves companies time and flexibility.
PaaS (Platform as a Service) is a cloud computing service that outsources the development platform, freeing developers from maintenance tasks to focus on their core business.
Artificial intelligence (AI) is becoming increasingly linked to the digital technologies we use every day, and has consistently improved their effectiveness. AI assimilates a lot of information with the data it receives. It then analyses this data to deliver the best possible action.
IaaS (« Infrastructure as a Service »), PaaS (« Platform as a Service ») et SaaS (« Software as a Service ») sont des services de cloud computing permettant aux entreprises d'accéder à des ressources en ligne à la demande, sans investissement en infrastructure.
Les agents d’IA révolutionnent la technologie en permettant des actions autonomes et un comportement axé sur des objectifs. Ils sont utilisés dans diverses applications, des chatbots aux voitures autonomes. OVHcloud propose des solutions pour soutenir le développement et le déploiement de l’IA.
Les superordinateurs sont des ordinateurs extrêmement rapides et puissants qui peuvent résoudre des problèmes complexes que les ordinateurs ordinaires ne peuvent pas gérer. Ils sont utilisés dans divers domaines, tels que la science, l'ingénierie et les affaires, pour la recherche, la conception et l'analyse de données.
Le format JPEG est omniprésent dans le monde numérique. Que ce soit pour partager des photos en ligne, afficher des images sur un site internet ou archiver des souvenirs précieux, ce format d'image est devenu une référence incontournable.
Big data
Machine learning is a sub-field of artificial intelligence (AI). It focuses on the development of algorithms and statistical models that allow a computer system to perform tasks without specific programming for them. In other words, instead of coding each instruction, we allow the machine to learn from information and improve its performance over time.
Discover machine learning, an innovative technology that turns big data into valuable insights for businesses.
Data mining refers to the collection, analysis and comparison of data from different and often unconnected sources. The data extracted can be used for many purposes, like improving efficiency and predicting behaviour, for example.
Deep learning is a machine learning method through which the machine is able to learn and improve on its own. This technique is based on artificial neurones, image recognition technologies and language recognition.
Blockchain is a decentralised ledger system that creates a secure and immutable record of transactions. Data is stored on a distributed network of computers called nodes, which are linked together in a chain and secured using cryptography.
By combining blockchain tech and peer-to-peer networking, Web3 allows you to wrestle the ownership of your data out of the hands of big tech and put it back into your own. This drives up privacy and security rates massively, all while reestablishing much-needed trust and independence in the online space.
A data lakehouse is an architectural approach combining features of data lakes and data warehouses. It supports both unstructured and structured data, offering vast storage (like a lake) and structured querying capabilities (like a warehouse). This architecture enables advanced analytics, AI, and ML, efficiently handling diverse data types at scale.
Machine learning, a subset of artificial intelligence, has an enormous impact on our everyday lives and tasks, and its role is increasing as the enterprise network becomes more sophisticated. Machine learning is defined as the process of a computer solving problems using algorithms.
Apache Kafka is an open-source streaming platform that facilitates real-time data transfer between different systems and applications. Built to be highly scalable and distributed, this event-streaming platform allows businesses to process, store, and transmit data streams efficiently. Kafka is commonly used for real-time analytics, data integration, and event-driven architectures.
Data Warehousing acts as a hub, integrating data from various sources into a single repository for analysis. It enables organizations to perform complex queries, uncover trends, and gain insights into operations, driving strategic decision-making through advanced analytics.
ETL (Extract, Transform, Load) is a critical process in data integration, consolidating diverse data sources into a unified, high-quality dataset for analysis and decision-making. It's essential for modern business intelligence (BI) initiatives, enabling organizations to gain valuable insights, improve operational efficiency, and make informed strategic decisions.
Reinforcement learning (RL) is a machine learning technique where agents learn to make decisions by interacting with an environment, receiving rewards for positive actions. This trial-and-error approach allows RL agents to optimize their behaviour over time, leading to impressive achievements in robotics, game playing, healthcare, and finance.
Data analytics is the process of transforming raw data into actionable knowledge. It involves collecting, cleaning, and analysing data to find patterns, correlations, and hidden trends. Data analytics is vital for businesses as it drives informed decision-making, optimises operations, identifies market trends, and manages risks.
Data governance is a system of processes and policies for managing and protecting data. It ensures data quality, security, and compliance with regulations. It also enables organizations to make better decisions, improve operational efficiency, and gain a competitive advantage.
Data warehouses consolidate diverse data sources, offering a single source of truth. They empower businesses with historical and current data for analysis, reporting, and informed decision-making. ETL (Extract, Transform, Load) processes prepare data for the warehouse, while BI tools unlock insights.
A data lake is a centralized storage space for storing raw, structured, semi-structured or unstructured data. It offers great flexibility for storing and analyzing this data, especially in big data and AI environments.
Data streaming is a process for processing data streams in real time. It enables continuous data analysis, providing better insights for quick decision-making. Data streaming is often used in areas such as social networks.
Le data lake stocke des données brutes et non structurées, tandis que le data warehouse organise des données structurées pour des analyses précises. Le choix dépend des besoins d'analyse.
Cloud computing
Explore our cloud computing solution, which provides on-demand resources. The Public Cloud is adapted to suit all projects, and offers a number of tools for containerisation, network, data analysis, and much more.
Load balancing is a technique where network traffic or user requests are distributed across multiple servers. This optimizes resource usage, improves performance, and ensures high service availability, avoiding overloads and failures.
Le stockage cloud est un type de stockage avec lequel les données sont stockées dans le cloud, sur des serveurs Internet gérés par un fournisseur de services cloud. Il s’agit d’une alternative flexible, évolutive et abordable au stockage sur site, qui permet aux organisations d’accéder à leurs données en ligne partout et à tout moment.
Cloud computing is a technology that provides computing resources (servers, data storage, network, etc.) on the Internet. Typically offered through a pay-as-you-go model, cloud computing offers a scalable and cost-effective alternative to on-premises (on-premise) infrastructures.
Hybrid cloud is an infrastructure that combines the use of public, private and dedicated cloud servers. Solutions are connected to communicate and manage applications and workloads securely, and with optimal performance.
Edge computing is a network architecture where data is processed very close to where it is generated and often far away from the core of the network. Since the processing is performed at the edge, latency is reduced as a result. This is a commonly used method for real-time workload processing.
Learn about cloud hosting and how it transforms data storage and website hosting for your business.
Quantum computing has the potential to change the game while revolutionising the computing landscape. While quantum computing can quickly solve complex problems that conventional computing cannot effectively handle, quantum computing is far from meeting expectations and many challenges remain.
A private cloud provides dedicated IT resources to a single user, delivering high levels of resource control, customisation, and increased security and compliance for an organisation’s data. It offers users the key benefits of the cloud, including scalability, elasticity and automation.
A cloud server is a computer accessible via the internet that corresponds to user requests and provides them with services such as websites, applications and platforms. It is managed by a user interface and maintained by a cloud service provider.
What is cloud migration and what are the benefits? Discover the 6 key strategies for a successful digital transformation with OVHcloud!
Anything as a Service (XaaS) enables enterprise to leverage a wide range of cloud computing tools and tech via a pay-as-you-go model, removing the need to install software or hardware or be responsible for its maintenance. This enables organisations to access next gen functionality almost overnight with zero friction.
Object storage takes huge volumes of unstructured data and turns each data point into a distinct unit – an object – with its own metadata and unique identifier. These objects are stored in a data environment, making it easier to access, retrieve and manage each and every individual object.
Containerisation has revolutionised how applications are developed and managed, offering a more flexible, portable and efficient way of managing them across various environments and infrastructure.
Cloud native lets business construct, deploy and maintain applications in the cloud within a flexible architecture. Crucially, these apps allow enterprise to leverage the many advantages offered by cloud-based services including the flexibility, scalability and resilience needed to remain competitive in tough market conditions.
Data storage ensures the safekeeping and accessibility of information using specialised technology. It involves on-premise storage with physical servers in a company's facility, and cloud storage, which utilises remote servers for scalable, accessible data management.
Container management (Conte refers to a set of practices that govern and maintain containerization software.
Cloud computing is a technology that provides access to a network of remote servers collectively called “the Cloud”. This means you can access and store your data, run programs, and use a variety of services from anywhere in the world, making your digital tasks more flexible and convenient.
Lift and shift sees enterprise migrating on-prem apps or workloads – and any related data – into the cloud without needing to redesign the solution. This enables a company to begin modernising its processes and workflows at speed as well as benefit from improved performance and reduced costs.
Grid computing is a powerful distributed computing model, connecting multiple computers to process large-scale tasks. It's integral to fields like science and finance, solving grand challenges through parallel processing. This technology is characterized by its use of diverse computers that are geographically dispersed, and co-ordinated through sophisticated middleware.
The main benefits of cloud computing for businesses are on-demand access to a wide range of IT services, provider-managed infrastructure maintenance, and the ability to develop private or hybrid clouds to suit their security and flexibility needs.
Block storage is a type of storage solution that stores data in equal-sized blocks, either on hard drives or in the cloud. This enables faster access to data compared to other storage methods, with the ability to increase capacity by adding more disks as and when you need them.
Cloud containers are lightweight, executable packages encapsulating everything needed to run software, including code, runtime, and system tools. Cloud containers offer a portable solution that ensures consistency across diverse environments from personal devices to public clouds. Through OS virtualization, containers isolate applications, providing each with dedicated resources.
A virtual machine (VM) is a computing resource that uses software rather than a physical computer to run programs and deploy applications, providing increased flexibility and efficiency in managing computing resources.
A virtual private cloud (VPC) is a service delivery model where a company is using an isolated section of a public cloud infrastructure. This allows for control and customization of the environment, ensuring data confidentiality and flexible management of IT resources.
Kubernetes is an open-source container orchestration platform that facilitates the deployment, scaling, and management of containerised applications in distributed environments. It automates deployment, resource management, and high availability tasks, providing a robust infrastructure for modern applications.
Infrastructure as Code (IaC) is a modern approach to managing computing infrastructure using code. It automates setup, configuration, and management, enabling faster, more reliable deployments. IaC increases speed and efficiency, enhances consistency and reliability, and promotes scalability and flexibility.
Natural Language Processing (NLP) is a specialised field of artificial intelligence that enables machines to understand, interpret, and generate human language. Using computational and linguistic techniques, NLP facilitates applications like virtual assistants, text analytics, and machine translation, transforming vast amounts of unstructured data into meaningful insights.
Cloud SDN (Software-Defined Networking) is a technology that manages network services through software, enhancing flexibility and scalability in cloud environments. It centralises control, optimises resource allocation, and simplifies network configuration and management, making it ideal for dynamic, virtualised cloud computing scenarios.
Containerization is a game-changing technology revolutionizing software development and deployment. Packaging applications and their dependencies into isolated units called containers offer numerous benefits, including portability, efficiency, scalability, and faster deployment.
Kubernetes is a powerful container orchestration engine that is ideal for complex applications but requires management expertise. Rancher simplifies Kubernetes adoption with a user-friendly interface and centralized multi-cluster management, making it accessible to a broader audience.
Cloud bursting allows an application to switch automatically between a private cloud and a public cloud to manage peaks in demand. When more capacity is needed, the application deploys the excess load on the public cloud to maintain performance.
Container orchestration is an automated process that manages the deployment, management, and scaling of containerised applications. It optimizes resource usage, ensures resilience and system scalability, making it easy to manage thousands of containers in complex environments like the cloud.
What is cloud migration and what are the benefits? Discover the 6 key strategies for a successful digital transformation with OVHcloud!
Explore our cloud computing solution, which provides on-demand resources. The Public Cloud is adapted to suit all projects, and offers a number of tools for containerisation, network, data analysis, and much more.
Load balancing is a technique where network traffic or user requests are distributed across multiple servers. This optimizes resource usage, improves performance, and ensures high service availability, avoiding overloads and failures.
Cloud computing is a technology that provides computing resources (servers, data storage, network, etc.) on the Internet. Typically offered through a pay-as-you-go model, cloud computing offers a scalable and cost-effective alternative to on-premises (on-premise) infrastructures.
A cloud server is a computer accessible via the internet that corresponds to user requests and provides them with services such as websites, applications and platforms. It is managed by a user interface and maintained by a cloud service provider.
Connected objects
The Internet of Things (IoT) represents all the devices or machines that form a network of internet-connected physical objects to transfer, collect and exchange data.
Automation is a term for technological applications in which the human role is reduced. Automation enables users to make simple, repetitive tasks automatic, and also automate processes that increase a company’s efficiency.
Computer vision is a field of AI that enables machines to interpret visual information like humans. It has diverse applications, from medical imaging and self-driving cars to manufacturing and augmented reality. Learning resources like OpenCV, TensorFlow, and online courses can help you get started.
Databases
With its range of features, PostgreSQL can process a range of database types with varying levels of complexity, while delivering optimal performance.
A relational database is a structure using SQL that provides access to data points, and in which information is linked through tables.
A cloud database is a database management system (DBMS) hosted on a cloud computing platform, be it a public, private, or hybrid cloud. Cloud databases share similar design and functionality with the on-premises databases traditionally found in an organization's data centre.
The primary distinction between on-premise and cloud databases lies in the deployment and management approach, but to end users and applications, cloud databases are indistinguishable from on-premises databases.
Data backup refers to the process of creating duplicate copies of digital data or files to safeguard against data loss if the originals are lost due to malicious action, accidents or disaster. The right backup processes will enable the enterprise to restore data quickly, causing minimal business disruption.
MongoDB is a source-available, cross-platform NoSQL database with optional schemas. It excels in storing large data sets, supporting various programming languages and features like ad hoc queries, indexing, and replication. Ideal for big data, web applications, and real-time analytics, MongoDB offers scalability, high performance, and a developer-friendly environment.
Any organisation typically holds a huge amount of data that must be effectively managed whether you’re saving, storing, retrieving or updating it. By introducing a data server, you get access to the back-end infrastructure and client-facing services you need to manage all your data efficiently via a single centralised ‘hub’.
A Large Language Model (LLM) is an advanced artificial intelligence system, specialised in natural language processing, capable of understanding, generating and interacting with human text by learning through vast textual data sets.
Learn about the unique characteristics of NoSQL databases, why they are so critical to information management today, and why they represent such an important development.
SQL is a powerful language designed for interacting with relational databases. Essential for both data analysis and supporting backend systems, it allows users to perform a range of tasks from basic data management to complex queries. With its wide adoption and ongoing evolution, SQL remains a fundamental skill in data-driven fields.
A database query allows you to ask a specific data-related question based on a pre-defined code. This is sent to a database management system (DBMS) with the system sourcing an answer direct from the database before retrieving and presenting it.
A non-relational database stores data in a flexible, unstructured way, unlike relational databases. It is suitable for non-homogeneous and scalable data, such as semi-structured or unstructured data, offering superior flexibility for data storage and retrieval.
A Database Management System (DBMS) is software that manages data within a database, offering features like data storage, retrieval, and manipulation. It bridges users with the database, ensuring data integrity, security, and accessibility, and handles vast data sets systematically.
Designed to be user-friendly, file storage saves, organises and manages data in a hierarchical structure – from files into folders and in turn, from folders into directories and subdirectories. This structure can be stored either on-prem or in the cloud via file storage hosting.
Data is perhaps the greatest single asset a business owns. It’s why processing data is so crucial to the continuing success and growth of any organisation.
From creation and storage to sharing and deletion, data lifecycle management (DLM) offers a framework for managing data effectively through out its life. Dedicated products automate each stage required by a DLM framework, helping business keep data secure, accurate and available 365 days a year.
Explore the world of databases with our concise guide. Learn when to use relational databases for structured data integrity and non-relational databases for flexibility in handling big data. Perfect for developers and IT managers.
SQL databases are relational and use SQL to manage data. NoSQL databases are non-relational and store data in flexible schemas. SQL is ideal for complex queries, while NoSQL offers more scalability for big data applications.
At the heart of quantum computing lies the qubit, a marvel capable of existing in a superposed state of 0 and 1. This unique characteristic empowers quantum computers to simultaneously explore vast computational possibilities, from cryptography to material science.
MLOps, or Machine Learning Operations, revolutionises the deployment and maintenance of ML models by merging DevOps principles with the machine learning lifecycle. It encompasses training, testing, deploying, and monitoring, aiming to automate these processes for improved reliability and efficiency.
MySQL is a popular open-source relational database management system that uses SQL (Structured Query Language) for managing, manipulating, and retrieving data. Known for its speed, reliability, and ease of use, MySQL supports various applications, from web development to big data projects, making it a versatile choice for developers and businesses.
DBaaS is a cloud computing model that eliminates the need for on-premises database management. It offers scalability, cost-effectiveness, and reduced management overhead. DBaaS is used in various applications, such as web development, IoT, gaming, and AI.
Domains & Web Hosting
Looking to create a website? Before it goes live, many elements need to be put in place first, including registering a domain name. But what does this involve?
A web hosting plan allows individuals or companies to store web data and publish websites online. More often than not, it’s a flexible service, offered by a hosting provider that offers a range of hosting options to companies of all sizes.
A bot is a software application that can be programmed to conduct many different tasks. They are fast and automated and are ideal for carrying out jobs that are repetitive, meaning bots are quicker at busywork than a human whose skills could be deployed elsewhere.
A top level domain (.TLD) is the last segment after the final full stop in a URL address, for example, .com or .org. A TLD is used to classify a website whether it’s commercial, charity, location-based or another specific variant.
A country code top-level domain (ccTLD) enables organisations to ‘label’ website addresses with a country-specific two-character identifier code such as, for example, .ca for Canada or .asia for Asia. This helps search engine rankings within specific locales while promoting a positive brand image to the locale’s citizens.
Standing for generic top-level domain, a gTLD is the last segment of a URL address after the full stop. Using popular extensions such as .com, gTLD can be deployed to promote general types of websites, services or products – and, unlike ccTLDs, are not associated with any specific country or locale.
A web application is software that is accessed and run through a website via the user’s browser. Businesses use web apps to deliver a wide range of in-browser features to customers, so there’s zero need for software downloads and installations.
With more and more websites coming online every year, cybersquatting has become a growing problem on the internet. This involves using or registering a domain name similar or identical to that of a well-known company.
A web server is a computer system combined with software used for the storing and serving of website files. Every time a user goes to a website, their client – a browser – requests the web page’s contents from the web server. In turn, the web site server delivers this to the user’s client.
Dedicated server hosting offers a physical server that is configured and used exclusively by a business, instead of being shared with other enterprise. This leads to performance, reliability and privacy gains across the board as well as giving the organisation complete control of its server solution.
In a computer network, localhost refers to the computer you are using to send a loopback request. This request is data from your computer acting as a virtual server, which is sent back to itself without ever passing through a physical network.
A Secure Sockets Layer (SSL) creates an encrypted link between a browser and a web server so any data moving between the two is kept private. The SSL protocol is actually the predecessor to the current protocol, TLS (Transport Layer Security), but SSL remains better known and its name remains in use today.
JavaScript is a browser-based programming language that brings interactivity to static web pages, applications and more. From animations to content updates, JavaScript is one of the three core web technologies, working alongside HTML and CCS to deliver dynamic web experiences.
Also referred to as a web address, a uniform resource locator (URL) is made up of different elements – such as a domain name – that instructs the web browser where and how it can access the desired resource that the user wants to view or retrieve.
HTML (Hypertext Markup Language) is a simple markup coding language used to layout content on a webpage. You use HTML programming language to tell web browsers how to display text, headings, images and other multimedia in the layout you want your users to see it in.
A domain name registrar is a company authorised to register domain names for websites, ensuring that an individual or a company can secure a unique website address. A generic top-level domain registry or a country code top-level domain registry must accredit registrars.
DNS attacks exploit the Domain Name System, crucial for internet functionality. Attackers manipulate DNS for unauthorised access, causing network disruptions, financial loss, and reputational damage. In cloud hosting, DNS security is a shared responsibility, crucial for protecting against data theft and malware
Within a business context, blog is website or page that is regularly updated with fresh posts that inform, educate, entertain and attract leads and customers, all while promoting the company’s online presence, thought leadership and brand.
Websites provide information, services, or products to users, and can serve various purposes, such as personal expression, business, education, or entertainment, accessible via the internet through a web browser.
Shared hosting is a web hosting service where multiple websites share the same server resources, making it a cost-effective solution for small businesses and personal websites. It’s ideal for those seeking an affordable, beginner-friendly hosting option with basic resource needs.
WordPress is a free, open-source content management system (CMS) used to build and manage websites and blogs. This page explains what WordPress is, its key features, benefits, and why it’s the most popular choice for web development.
Nginx is a high-performance web server that excels at handling high traffic volumes with its efficient event-driven architecture. Versatile in roles like reverse proxying, load balancing, and caching, it significantly boosts website speed and security. Nginx's flexible configuration and lightweight design make it popular for diverse web applications, from simple blogs to complex enterprise solutions.
JSON, a lightweight data interchange format, simplifies how computers and applications communicate. Its structure, using key-value pairs and arrays, facilitates data organisation and enables seamless interaction across different platforms.
Static websites deliver pre-built HTML files directly to users, resulting in exceptional speed and security. They're perfect for blogs, portfolios, and informational sites where content updates are infrequent. While they lack the dynamic features of database-driven sites, their performance and ease of management make them an attractive option for many projects.
L’empoisonnement des DNS, également appelé « empoisonnement du cache DNS », compromet l’intégrité des DNS (Domain Name System) en redirigeant les utilisateurs vers des sites web malveillants. Cela peut conduire au vol d'informations sensibles ou à l'installation de logiciels malveillants sur les appareils des victimes sans méfiance.
Infrastructure
A datacentre can be described as a building and/or infrastructure that hosts many computers. For example, they can be used to store a company’s IT system data. As simple as this definition seems, the technologies and systems in place are very complex. Take one of OVHcloud’s datacentres for example, that operates across the globe.
Dedicated servers are physical computers located in a datacentre, with all servers’ resources allocated exclusively to a single client. Dedicated servers are powerful, highly flexible and used as the foundation for many complex IT projects.
Bare metal servers are physical services based on a server and allocated to an individual customer. Each server is a physical device without a hypervisor preinstalled, and all of its resources are dedicated to the single client.
A teraflop corresponds to a computer’s speed in performing one trillion floating-point operations per second. It is used to measure computing performance, and is particularly crucial in supercomputers, video games and complex data processing.
Discover DevOps: a synergy of development and operations boosting efficiency and speed in software delivery processes.
IT infrastructure refers to the combined group of technology elements including hardware, software, networks, facilities, and all the necessary components that support the management, processing, and storage of information.
HCI revolutionises IT by merging compute, storage, and networking into a single system. With software-defined elements and a focus on virtualization, HCI offers a streamlined, efficient approach to management. It offers flexibility, cost savings, and simplified scaling. Embrace HCI for cloud transformation, data centre modernization, and a unified management experience.
High Performance Computing (HPC) harnesses supercomputers and clusters for complex, high-speed computations. Essential in scientific, engineering, and business domains, HPC performs quadrillions of calculations per second, vastly outpacing standard computers. HPC enables rapid processing of vast data sets and simulations, supporting advancements in fields including healthcare, aviation, and manufacturing.
Server monitoring is the process of systematically tracking, measuring and observing the operations and processes on a server. The primary goal of server monitoring is to collect data that can be used to assess the health and condition of the server, ensuring that it operates at optimal performance levels and ensuring applications such as software-as-a-service (SaaS) run flawlessly.
Nutanix simplifies IT infrastructure by converging compute, storage, and networking. It enhances performance, supports cloud-native applications, and improves efficiency while strengthening data security. Nutanix is a powerful solution for businesses seeking a modern, flexible, and secure IT infrastructure.
Hypervisors are software, firmware, or hardware that create and manage virtual machines, enabling multiple operating systems to run on a single physical machine. They optimize resource utilization, enhance security, and improve scalability, leading to cost savings and increased efficiency in IT infrastructure.
IOPS, or Input/Output Operations Per Second, measures storage device performance, particularly for applications with many small, random data requests. SSDs offer higher IOPS than HDDs, making them ideal for demanding tasks. Understanding your IOPS requirements is key to choosing the right storage solution.
Virtualisation
A virtual private server (VPS) is a virtual server created from a physical machine. The server’s resources are fully allocated to you, guaranteeing total freedom in terms of how you choose to configure it. This technology can also be adapted to suit your projects.
Server virtualization technology enables the creation of multiple virtual servers from a single physical server. It helps organizations to cut costs, drive efficiency, and scale easily.
Server virtualization technology enables the creation of multiple virtual servers from a single physical server. It helps organizations to cut costs, drive efficiency, and scale easily.
A private cloud provides dedicated IT resources to a single user, delivering high levels of resource control, customisation, and increased security and compliance for an organisation’s data. It offers users the key benefits of the cloud, including scalability, elasticity and automation.
VDI offers remote employees access to their desktop environment from anywhere, fostering productivity and flexibility. By centralizing data and applications on remote servers, VDI enhances security, reducing risks associated with lost or stolen devices. With centralized management, IT teams can efficiently handle updates and security patches, simplifying administrative tasks and potentially reducing costs.
Security and compliance
Data encryption encompasses all methods of encrypting sensitive or personal information in order to ensure its confidentiality. The data can only be read with an encryption key.
Data security encompasses all of the best practices in terms of protection that must be put in place to guarantee the confidentiality, sovereignty and integrity of the data stored on an IT infrastructure.
Cybersecurity is the application of processes and technologies to protect computer systems, networks and data from cyberattacks, unauthorised access and leaks. It helps maintain data availability, integrity and confidentiality.
GRC stands for Governance, Risk, and Compliance, acting as a framework within IT services and security. It aligns IT efforts with organisational goals, ensuring security and compliance with industry standards while managing IT risks like cyber threats and data breaches.
A private network is secured and isolated from the public internet, allowing only authorised devices and apps to connect and exchange data with one another. Private networks are often used by business to create local area networks, which offer secure access to network resources within a defined geographic location.
Disaster recovery is an organisation’s strategy for protecting its data and ensuring business as usual after a disruptive event. Learn more here.
Identity Access Management offers a framework of technologies and processes that only allow users with official authorisation to access enterprise data, resources, assets, and more. This helps protect business from malicious actions and ensures best IT practices are maintained, all while enabling regulatory compliancy targets to be met.
Serverless computing is transforming how applications are built and deployed by abstracting infrastructure management. Developers write code while cloud providers handle scaling, provisioning, and security. This approach enables faster development, cost efficiency, and increased focus on innovation.
DevSecOps seamlessly integrates security into every stage of software development. It fosters collaboration between teams and automates security checks, enabling faster and more secure software releases.
Customer Identity and Access Management (CIAM) solutions secure and enhance customer experiences. They streamline registration, protect data, offer personalisation, and drive engagement. Businesses can build loyalty and gain a competitive edge by understanding preferences and behaviours.
In a world where data security and online privacy have become major concerns, VPNs are proving to be essential tools. Find out how they meet the increasing demands of protection and flexible access in an increasingly complex digital landscape.
Confidential computing safeguards sensitive information, even in use, through advanced hardware-based isolation and encryption. It enhances data privacy, ensures compliance, and builds trust in the cloud. Expert solutions empower you to innovate while maintaining the highest level of protection.
Cloud security is a group of measures and technologies that protect data, applications and infrastructure in IaaS, PaaS and SaaS services. It prevents unauthorized access, manages incidents, and ensures compliance.