


Three best practices for small and medium-sized enterprises to adopt hybrid cloud to handle big data
Today, big data and analytics are entering a more mature deployment stage. This is good news for small and medium-sized businesses that are deploying these technologies and have been struggling to define a big data architecture for their company.
Uncertainty about how to define the overall architecture of big data and analytics is one of the reasons why SMBs lag behind in big data and analytics deployment. In many cases, they are waiting and watching to see how trends such as hybrid computing, data marts, master databases, etc. develop, and how controls over security and governance will play out.
Finally, an emerging best practice data architecture that everyone can follow will be provided. In this architecture: Cloud computing services are being used to store and process big data, while on-premise data centers are used to develop local data marts in the enterprise.
Let’s take a closer look at the reasons behind this big data and analytics architecture:
If the enterprise is small, it is expensive to purchase server clusters to process big data in parallel in the data center, not to mention hiring or training very expensive professionals who know how to optimize, upgrade and maintain the parallel processing environment. . Businesses that choose to process and store data on-site also make significant investments in hardware, software, and storage equipment. Procuring big data hardware and software, and outsourcing computing processing and storage to the cloud will all cost a lot of money.
Data governance (for example, security and compliance issues) is one of the reasons why enterprises are reluctant to deliver all their mission-critical data to the cloud because it is more difficult to manage this cloud data. Therefore, once the data is processed in the cloud, many enterprises choose to migrate the data to their own on-premises data centers.
There is another reason why many enterprises choose to use their data centers: to focus on the proprietary applications and algorithms that develop this data, because it is the policy of many cloud computing providers that any applications developed by customers in the cloud may be compared with other Customer sharing.
By keeping applications on-premises in the data center and developing an on-premises master data set from which smaller data marts can be separated, enterprises have direct control over their data and applications.
For example, if an enterprise needs to anonymize data, the process it implements should be documented and agreed with its cloud computing provider, as the cloud computing provider will perform the anonymization. If an enterprise wants to clean up its own data, it should also provide detailed written instructions to its cloud computing provider on the cleanup process. For example, does the business just want to unify the abbreviations for all U.S. states (e.g., "Tenn" and "Tennessee" = "TN") or do other edits to the data make it uniform and easier to process? In the end, whether your business is Whether running in a cloud computing service provider's dedicated tenant or in a multi-tenant environment, the cloud computing provider should be able to guarantee that the enterprise's data is never shared with other customers.
Many IT departments in enterprises completely miss this task. They just start implementing big data projects but forget that the existing application development policies and procedures come from the application domain of the transaction. Businesses should not make this mistake. Instead, companies need to revise policies and procedures in areas where the likelihood of interacting with big data is higher (e.g., storage, database management, applications).
In the case of cloud-based disaster recovery (DR) testing, enterprises should include provisions in the contract for documenting and executing DR. Disaster recovery (DR) plans (which focus on transactional data and systems) should also be kept up to date and include recovery and test scripts for big data and analytics.
The above is the detailed content of Three best practices for small and medium-sized enterprises to adopt hybrid cloud to handle big data. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

Video Face Swap
Swap faces in any video effortlessly with our completely free AI face swap tool!

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics











The five basic components of the Linux system are: 1. Kernel, 2. System library, 3. System utilities, 4. Graphical user interface, 5. Applications. The kernel manages hardware resources, the system library provides precompiled functions, system utilities are used for system management, the GUI provides visual interaction, and applications use these components to implement functions.

vscode built-in terminal is a development tool that allows running commands and scripts within the editor to simplify the development process. How to use vscode terminal: Open the terminal with the shortcut key (Ctrl/Cmd). Enter a command or run the script. Use hotkeys (such as Ctrl L to clear the terminal). Change the working directory (such as the cd command). Advanced features include debug mode, automatic code snippet completion, and interactive command history.

To view the Git repository address, perform the following steps: 1. Open the command line and navigate to the repository directory; 2. Run the "git remote -v" command; 3. View the repository name in the output and its corresponding address.

Although Notepad cannot run Java code directly, it can be achieved by using other tools: using the command line compiler (javac) to generate a bytecode file (filename.class). Use the Java interpreter (java) to interpret bytecode, execute the code, and output the result.

The main uses of Linux include: 1. Server operating system, 2. Embedded system, 3. Desktop operating system, 4. Development and testing environment. Linux excels in these areas, providing stability, security and efficient development tools.

VS Code One-step/Next step shortcut key usage: One-step (backward): Windows/Linux: Ctrl ←; macOS: Cmd ←Next step (forward): Windows/Linux: Ctrl →; macOS: Cmd →

Causes and solutions for the VS Code terminal commands not available: The necessary tools are not installed (Windows: WSL; macOS: Xcode command line tools) Path configuration is wrong (add executable files to PATH environment variables) Permission issues (run VS Code as administrator) Firewall or proxy restrictions (check settings, unrestrictions) Terminal settings are incorrect (enable use of external terminals) VS Code installation is corrupt (reinstall or update) Terminal configuration is incompatible (try different terminal types or commands) Specific environment variables are missing (set necessary environment variables)

There are six ways to run code in Sublime: through hotkeys, menus, build systems, command lines, set default build systems, and custom build commands, and run individual files/projects by right-clicking on projects/files. The build system availability depends on the installation of Sublime Text.
