【24h】

Market Analysis: A Bigdata Solution

机译:市场分析:大数据解决方案

获取原文
获取原文并翻译 | 示例
           

摘要

In this paper we present the Bigdata infrastructure for handling the large amounts of data processed in the Investment industry. Various bigdata tools can be used for this purpose. But in this case, we are going with Hadoop. Hadoop is framework which is developed using Java programming language. It is a framework which uses various concepts of parallel and distributed computing to make the computational speed faster. This causes the programs to execute at a much larger speed with the help of few normal speed computers. This increases the affordability rate and makes it much more efficient. It uses its own file system called as Hadoop distributed file system, that is HDFS. HDFS is known for its security and high risk control. Since it runs in cluster there becomes absolutely no use of a Super computer to process data faster. Hadoop is the widely used big data processing engine with a simple master slave setup. One of the most common place where bigdata is most commonly uses is the share market industry. There are various reasons why bigdata is used in this field. The most common one being to increase the profits by understanding the pervious data. The analytics and understanding of data can only be possible if the large amounts of data is handled in a proper way. Suggesting the shares to users is one of the main concept of this paper. But rather than focusing on the analytical part of the framework our main aim is to make it easier for the admin to use bigdata so that the large amounts of data sets can be easier to process. This application can have a lot of advantages in the algorithmic trading. Algorithmic trading is a type of trading where different algorithms are used for buying and selling the shares. It can also have various use cases in stock brokering firms for processing large amount of data quickly. The main contribution of this paper is to integrate the cloud computing and Hadoop framework. The cloud computing in this project is a web based application which is directly connected to the Hadoop system. The parallel pipeline is developed for the purpose of easy handling of data by the admin. We also have developed a communication protocol upon TCP/IP for the purpose of the pipeline. The share market industry involves are huge amount of data and exabytes of data is processed every minute for various purposes. The investors usually go through all the data that is involved in their research purposes and try to analyse it. There are various factors and also attributes that the investors try to take into account when going through the data. Analysing data involves a large amount of data and very well built platform for it to support the data. In this paper, we have also built a web based platform which helps user analyse the data in a much simpler and graphical notation. The attributes include financial transaction patterns by the investor, market conditions and sentiments, macroeconomics variables, scheme level features, and demographic factors. Predicting the redemption behavior requires sophisticated platform that can capture multiple factors that affect the redemption behavior. However, these big data infrastructure provide us with various use cases, with tools like Hadoop and spark it becomes even more easier to find use cases at macro levels. This platform can investigate these factors for near real-time data and can provide highly accurate predictions for the redeeming investors in the future at an investor-level. Our results show that by implementing cloud computing, bigdata analytics and sofisticated algorithmic trading the results which are data driven can be used to generate a resonable amounts of profits and also the data could be processed in a much simple and faster way.
机译:在本文中,我们提出了处理投资业中处理的大量数据的基础设施。各种BIGDATA工具可用于此目的。但在这种情况下,我们与Hadoop一起去。 Hadoop是使用Java编程语言开发的框架。它是一种使用并行和分布式计算的各种概念的框架,使计算速度更快。这使得程序在很少的正常速度计算机的帮助下以更大的速度执行更大的速度。这增加了可负担性率,使其更有效。它使用自己的文件系统称为Hadoop分布式文件系统,即HDFS。 HDFS以其安全性和高风险控制而闻名。由于它在群集中运行,因此绝对不会使用超级计算机以更快地处理数据。 Hadoop是广泛使用的大数据处理引擎,具有简单的主从设置。大数据最常用的最常见地点之一是股票市场行业。在此领域中使用大数据的原因有很多原因。最常见的是通过了解透水数据来增加利润。如果以适当的方式处理大量数据,则才能进行分析和对数据的理解。向用户建议股票是本文的主要概念之一。但不是专注于框架的分析部分我们的主要目的是使管理员更容易使用BigData,以便更容易处理大量数据集。此应用程序可以在算法交易中具有很多优势。算法交易是一种交易类型,其中不同的算法用于购买和销售股票。它还可以在股票交易公司中拥有各种用例,用于快速处理大量数据。本文的主要贡献是集云计算和Hadoop框架集成。该项目中的云计算是基于Web的应用程序,它直接连接到Hadoop系统。并行管道是为了易于处理管理员的数据而开发的。我们还在TCP / IP上开发了一种用于管道的通信协议。股票市场行业涉及的是大量数据,每分钟处理数据的exabytes以各种目的。投资者通常经过研究目的所涉及的所有数据,并尝试分析它。有各种因素,也有归因于通过数据何时尝试考虑到帐户。分析数据涉及大量数据和非常良好的构建平台,以支持数据。在本文中,我们还建立了一个基于Web的平台,帮助用户以更简单和图形表示法分析数据。该属性包括投资者,市场条件和情绪,宏观经济变量,计划级别特征和人口因子的金融交易模式。预测兑换行为需要复杂的平台,可以捕获影响赎回行为的多个因素。然而,这些大数据基础架构为我们提供各种用例,具有像Hadoop和Spark这样的工具,它变得更加容易在宏观上找到用例。该平台可以研究近实时数据的这些因素,可以在投资者级别为未来提供高度准确的预测。我们的研究结果表明,通过实施云计算,BigData分析和Soficated算法交易是数据驱动的结果可用于生成可谐振量的利润,并且还可以以更简单且更快的方式处理数据。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
获取原文

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号