DATA SCIENCE

AtScale Announced a Partnership with Datarobot to Drive Collaboration between Data Science and Business Intelligence Teams

AtScale | February 03, 2021

AtScale, the main supplier of wise information virtualization, today announced a partnership with DataRobot, the main undertaking AI stage, to convey a turnkey way to deal with prescient and enlightening information examination. The partnership constructs an extension between information science and business insight to fuse and have a similar KPIs utilized for prescient and prescriptive business choices.

“We are thrilled to partner with DataRobot to accelerate our go-to-market within the data science community. DataRobot’s vision for maximizing business value via AI is unparalleled, and the combination of our capabilities is another step toward democratization of predictive and prescriptive analytics,” said Christopher Lynch, executive chairman and CEO, AtScale.

Associations worldwide are moving AI forecasts into everyday business tasks. The mix of DataRobot and AtScale furnishes a consumable venture measurements center with steady, administered KPIs and a streamlined interface for computerized highlight creation using DataRobot's Feature Importance and Impact usefulness. AtScale extraordinarily makes DataRobot expectations consumable to business insight and detailing workstreams, giving both specialized and non-specialized partners admittance to forecasts utilizing their preferred apparatus.

“Our global market presence and diverse customer base continue to highlight the need for greater unity between the data science and business intelligence communities,” said Jeremy Achin, co-founder and CEO, DataRobot. “AtScale’s intelligent data virtualization adds a new business constituency in the enterprise who can incorporate AI predictions for operational decision making.”

About AtScale

AtScale powers the analysis used by the Global 2000 to make million dollar business decisions. The company’s Intelligent Data Virtualization platform provides organizations the opportunity to enhance or create a self service data culture by leveraging its business-friendly semantic layer, intelligent data engineering, easy-to-use web-based design interface, multi-cloud support, and strong data source and client integration.

About DataRobot

DataRobot is the leader in enterprise AI, delivering trusted AI technology and enablement services to global enterprises competing in today’s Intelligence Revolution. DataRobot’s enterprise AI platform democratizes data science with end-to-end automation for building, deploying, and managing machine learning models. This platform maximizes business value by delivering AI at scale and continuously optimizing performance over time. The company’s proven combination of cutting-edge software and world-class AI implementation, training, and support services, empowers any organization – regardless of size, industry, or resources – to drive better business outcomes with AI.

Spotlight

Hugh Owen speaks to World 2015 attendees about data visualization and discovery.


Other News
BIG DATA MANAGEMENT

Talend Acquires Gamma Soft, a Market Innovator in Change Data Capture

Talend | April 11, 2022

Talend, a global leader in data integration and management, announced today it has acquired Gamma Soft, a market innovator in change data capture (CDC). The addition of Gamma Soft's highly complementary, enterprise-class change data capture technologies will help customers streamline their data modernization initiatives, including cloud migrations, and support advanced, real-time analytics use cases across hybrid and multi-cloud environments. Today, many organizations rely on brittle, hand-coded integrations, or rely on multiple data management tools with redundant capabilities across integration, replication, modeling, preparation, quality, cataloging, and governance. With the combination of Talend and Gamma Soft, data professionals will be able to solve more use cases that require support for quickly changing data faster and easier than ever on a single end-to-end solution. "We are thrilled to welcome the talented Gamma Soft team to Talend. Complementary to our product portfolio, Gamma Soft deepens our already comprehensive integration capabilities and gives us new functionality for enabling advanced, real-time business insight. More broadly, Gamma Soft extends the value we provide customers in helping them quickly build, continually monitor, and easily optimize enterprise-wide data health." Christal Bemont, CEO, Talend Headquartered in Paris, France, Gamma Soft helps companies continuously track and replicate changed data in real time from a source, such as data warehouses, data lakes, and other databases, to a destination without requiring the entire data set to be extracted. This process provides multiple benefits, including streamlining and accelerating cloud data migration projects and enabling real-time business optics to drive everything from supply-chain optimization to fraud detection. "Change data capture technologies offer speed, accuracy, and agility in data replication that can help businesses successfully optimize their real-time analytics and cloud migration initiatives," said Stewart Bond, Research Director, IDC. "According to our recent market forecast, taking control of dynamic data is a high priority for companies that need to continue their digital transformation and plan for digital resiliency. Bringing Gamma Soft into Talend's product portfolio is a great add for Talend and for its customers." Véronique Goussard, general manager, Gamma Soft said, "Joining Talend is a great fit from a product and cultural perspective for Gamma Soft and for our customers. Talend will help take our CDC capabilities to the next level and provide customers with a single, end-to-end solution to successfully execute on data strategies that rely on quickly capturing changing data for analysis in cloud, hybrid or multi-cloud implementations." About Talend Talend, a leader in data integration and data management, is changing the way the world makes decisions. Talend Data Fabric is the only platform that seamlessly combines an extensive range of data integration and governance capabilities to actively manage the health of corporate information. This unified approach is unique and essential to delivering complete, clean, and uncompromised data in real-time to all employees. It has made it possible to create innovations like the Talend Trust Score™, an industry-first assessment that instantly quantifies the reliability of any data set.

Read More

DATA SCIENCE

Saturn Cloud and Bodo.ai Partner to Bring Extreme Performance Python to Data Scientists

Saturn Cloud | June 02, 2022

Saturn Cloud, the data science and machine learning platform and bodo.ai, a parallel data compute platform providing extreme scale and speed for Python, have announced their partnership to take Python analytics performance to the next level for data science teams. Data scientists develop multiple workflows across teams, and rely on Saturn Cloud to provide a collaborative environment and computing resources. With this partnership, those teams now have seamless access to the Bodo platform - allowing them to scale prototypes to petabyte-scale parallel-processing production without any tuning or re-coding. Saturn Cloud's pre-built tools allow data science teams to collaborate and scale easily, without locking users into patterns. Instead, the platform encourages the workflow the user already has, while providing an environment where they don't need to rely on dev sources or manage compute environments. It prioritizes keeping the data scientist self-sufficient, while being able to collaborate and share work more efficiently. Bodo offers a parallel compute platform providing extreme scale and speed, but with the simplicity and flexibility of using native Python. In contrast to using libraries and frameworks like Spark, Bodo is a new type of compiler offering automatic parallelism and high efficiency surpassing 10,000+ cores. Bodo can also be used natively with analytics packages such as Pandas, NumPy, SciKit Learn, and more. The joint solution is available immediately, with bodo.ai software running within Saturn Cloud resources. Saturn Cloud provides a pre-built template with Bodo already installed and configured. Then, users are able to access the functionality of bodo.ai within JupyterLab or via SSH from VSCode, PyCharm, or the terminal. By using Saturn Cloud, users are able to get up to 4TB of RAM and 128 vCPUs, all backing the powerful software of Bodo. You can try the following examples right away here: Use Bodo to speed up feature engineering and model training or use Bodo to speed up data manipulation and analysis. "Our partnership is focused on providing massive speed and productivity improvements to data scientists struggling with large-scale analytics projects. Bodo's platform adds terabyte-scale processing with unheard-of infrastructure efficiencies for Saturn Cloud users." Behzad Nasre, CEO, Bodo "We not only want to provide a flexible workspace for data science teams, but enable greater Python scaling capabilities to increase productivity in projects that are more demanding. This joint offering with Bodo will give users an opportunity to take their work to the next level with automatic parallelization for better overall performance," says Sebastian Metti, one of the Saturn Cloud founders. About Saturn Cloud Saturn Cloud is a data science and machine learning platform flexible enough for any team. Collaborate together in the cloud on analyses and model training, then deploy your code. All using the same patterns you're used to, but with cloud scale. Learn more here. About Bodo Founded in 2019, Bodo.ai is an extreme-performance parallel compute platform for data analytics, scaling past 10,000 cores and petabytes of data with unprecedented efficiency and linear scaling. Leveraging automatic parallelization and the first inferential compiler, Bodo is helping F500 customers solve some of the world's largest data analysis problems. And doing so in a fraction of traditional time, complexity, and cost, all while leveraging the simplicity and flexibility of native Python. Developers can deploy Bodo on any infrastructure, from a laptop to a public cloud.

Read More

DATA ARCHITECTURE

Veeva Data Cloud Brings Together OpenData, Link, and Compass on a Common Data Architecture

Veeva Systems | April 27, 2022

Veeva Systems today announced Veeva Data Cloud, a group of cloud data applications that includes Veeva OpenData customer reference data, Veeva Link for real-time intelligence, and Veeva Compass patient, prescriber, and sales data for the U.S. market. Veeva Data Cloud reflects the company's long-term commitment to delivering better data to the industry, all on a common data architecture. Taking a new approach, Veeva Data Cloud delivers unlimited access to data through modern cloud software for greater flexibility and data that's always up to date. "Commercial excellence is about getting your products to the patients that need them. With Veeva Data Cloud's modern approach and common data architecture, sales, medical, and marketing teams can work together in a more coordinated, customer-centric, and compliant way." Veeva CEO Peter Gassner Veeva Data Cloud's three major product families all share a common data architecture and "better data, better delivery" approach: Veeva OpenData is customer reference data, including healthcare provider (HCP) and healthcare organization data, and the affiliations between them. OpenData delivers high-quality data with no usage restrictions in more than 65 countries today, with plans for more than 100 countries by the end of 2023. Veeva Link data applications deliver real-time intelligence on key scientific and medical experts, health systems, scientific awareness and sentiment, and medical insights. Veeva Compass is a suite of U.S. longitudinal patient, longitudinal prescriber, and sales data for a wide range of commercial use cases, including launch planning, HCP segmentation and targeting, incentive compensation, and patient journey analytics. Veeva Compass Patient and Veeva Compass Prescriber are available today, with Veeva Compass Sales planned for availability in late 2022. Veeva Data Cloud and Veeva Commercial Cloud share a common data architecture so customers can work seamlessly across sales, medical, and marketing. This common data architecture includes shared definitions for data elements such as brands, doctors, products, content, and interactions. For more information, see the Veeva Summit keynotes on-demand at veeva.com/SummitKeynotes. Life sciences industry professionals can also register for the Veeva Commercial Summit for sales, medical, and marketing in Boston on May 5. About Veeva Systems Veeva is the global leader in cloud software for the life sciences industry. Committed to innovation, product excellence, and customer success, Veeva serves more than 1,000 customers, ranging from the world's largest pharmaceutical companies to emerging biotechs. As a Public Benefit Corporation, Veeva is committed to balancing the interests of all stakeholders, including customers, employees, shareholders, and the industries it serves.

Read More

BIG DATA MANAGEMENT

Penguin Releases the Decentralized Data Network for Web3.0

Penguin | January 03, 2022

Recently, Penguin team has announced the launch of their decentralized data network for Web3.0. With the advancement of blockchain technology, some innovative new players are entering the market. Some are bringing the offline world to a global audience, while others transform the way we invest in our future. Decentralized applications, DeFi, NFTs, and the Metaverse, hold immense potential for future growth and real-world uses. But what the current crypto arena lacks is an independent & one-stop web service that includes a high-performance smart contract blockchain together with a decentralized storage solution. The Penguin network brings in a universal decentralized data network specifically designed for Web 3.0. Penguin - The Decentralized Storage Platform Exclusively designed for Web 3.0, Penguin is a peer-to-peer network of nodes, which jointly provides decentralized storage and communication service. By offering a universal decentralized data network for Web3.0, the platform can fulfill multiple roles for different areas of blockchain space. Moreover, Penguin aims to work with the blockchain industry to create decentralized applications (DApps), products, and services seamlessly accessible in Web 3.0. A unique feature of the platform is that it offers automatic scaling; that is, an increase in storage space demand would be efficiently handled. This will eventually lead to a lowering of costs for the blockchain arena. Penguin also facilitates efficient data storage capabilities and quick data retrieval. The network is economically automated with a native protocol token, PEN, thanks to its built-in smart-contract-based incentive system. Therefore, the purported goal of the platform is to extend the blockchain by utilizing decentralized storage and communication to position itself as a world computer that can efficiently serve as an operating system and deployment environment for dApps. Web 3.0 - The Decentralized Internet of the Future Web 3.0 is not merely a buzzword that tech, crypto, and venture-capital classes have become interested in lately. It aims to provide a future where distributed users and machines can seamlessly interact with data, value, and other counterparties through peer-to-peer networks, eliminating the need for any third parties. It is built majorly on three novel layers of technological innovation. Those are edge computing, decentralized data networks, and artificial intelligence. Web 3.0, built on blockchain, eliminates all big intermediaries, including centralized governing bodies or repositories. Moreover, the most significant evolution enabled by Web 3.0 is the minimization of the trust required for coordination on a global scale. It fundamentally expands the scale and scope of human and machine interactions to a far new level. These interactions range from easy payments to richer information flows and trusted data transfers, all without passing through a fee-charging intermediary. Web 3.0 enhances the current internet service with significant characteristics like trustless, verifiable, permissionless, self-governing, etc. This is why a permissionless, decentralized blockchain like Penguin plays a pivotal part in developing the so-called "decentralized internet of the future." Decentralized data networks like Penguin make it possible for data generators to store or sell their data without losing ownership control, compromising privacy, or reliance on intermediaries or go-betweens. Blockchain Technology and Web 3.0 Blockchain technology and cryptocurrencies have always been an integral part of Web3.0. It provides financial incentives for anyone who wants to create, govern, contribute, or improve projects. Today the internet needs Web 3.0, a new generation of the Internet protocol that facilitates free identity, free contracts, and free assets. Blockchain technology with its advanced network fundamentals offers a near-perfect solution with in-built smart contracts for self-deployment and access, decentralized addresses as accounts, etc. Penguin, the decentralized data network, provides an available decentralized private data storage solution for all Web3.0 developers. How Does Penguin Benefit The Development Of Web 3.0 Today we live in a data-driven world, where companies often collect massive amounts of user data and use this data with the intent to deliver value. Data privacy has become a greater concern over the past few years. However, the Internet ecosystem has fundamentally changed several concerns like data privacy and storage. This is referred to as Web 3.0, and it ensures this by deploying blockchain. Penguin primarily focuses on data storage with zero downtime. It also features permanent versionable content storage, zero error operation, and resistance to intermittent disconnection of nodes. With its exceptional privacy attributes like anonymous browsing, deniable storage, untraceable messaging, and file representation formats that leak no metadata, Penguin meets with the growing security demand on the web. Penguin also offers continuous service and resilience against outages or targeted attacks. The platform facilitates the creation of many products, where all products rely on APIs and SDKs provided by Penguin. Penguin - An Infrastructure for A Self-Sovereign Society Penguin is more than just a network; the protocol sets a strong foundation for creating a market economy around data storage and retrieval. The platform also has entered into a host of prospective and strategic partnerships and collaborations with different projects and protocols in the DeFi, GameFi, NFTs, smart contract, and other metaverse spaces. Moreover, as a platform for permissionless publication, the Penguin network promotes information freedom. The platform’s design requirements can only be met by the network native token PEN. Some of the significant features that Web 3.0 offers are zero central point of control by removing intermediaries, complete ownership of data, sharing information in a permissionless manner, reducing hacks and data breaches with decentralized data, and interoperability. On the other hand, Penguin aims to build an infrastructure for a self-sovereign society. Without permission and privacy, Penguin efficiently meets the needs of freedom of speech, data sovereignty, open network market, and ensuring its security through integrity protection, censorship resistance, and attack resilience. Some of its vital meta values are Inclusivity, the need to include the underprivileged in the data economy, lowering the barrier of entry to explain complex data flows, and building decentralized applications. The integrity of the online persona is necessary. Because Penguin is a network with open participation and offers services and permissionless access to publishing, sharing, and investing your data, users have complete freedom to express their intention and have full authority to decide whether they want to remain anonymous or share interactions. Incentivization or economic incentives ensure that participants' behavior aligns with the network's desired emergent behavior. Finally, Impartiality guarantees content neutrality and prevents gate-keeping. It successfully rules out other values that treat any particular group as a privileged or express preference for specific content or even data from any specific source. These meta values make Penguin an efficient decentralized, permissionless data network for Web 3.0. Penguin’s Future-Proof Design Principles - Meeting the Needs of Web 3.0 The information society and data economy have ushered in an era where online transactions and big data are pivotal for everyday life. Therefore, it is essential to have a future-proof and advanced supporting technology like Penguin. The network offers a strong guarantee for continuity. The Penguin network ensures continuity by following some general requirements or system attributes. Some of them are stable and resilient specifications and software implementation. Scalable enough to accommodate many orders of magnitude, more users, and data without lowering the performance or reliability for mass adoption, secure and resilient solution to deliberate attacks, Penguin is a self-sustaining autonomous solution that is independent of human or organizational coordination or any legal entity's business.

Read More

Spotlight

Hugh Owen speaks to World 2015 attendees about data visualization and discovery.

Resources

Whitepaper

Whitepaper

Whitepaper