Git Product home page Git Product logo

scdp's Introduction

SCDP

Smart City Data Platform Graz

scdp's People

Contributors

anothercodeartist avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar

scdp's Issues

Data concentrator for dealing with a large number of values

Comet apparently has a strict limit on the number of values, on the other side browsers can only handle a certain maximum number of values.
We need an operator that reduces the number of entries in a list by replacing chunks of data with their average.
Idea: Config setting with maximum number of values. If the data collection exceeds this value, its reduces to about this given size
Update/Remark: Best implemented as additional feature within Comet. This would limit the data volume exchanged between Comet and Wirecloud

FIWARE Performance Test

Perform a series of performance tests on different container layouts in order to find the optimal layout for FIWARE components

Implement a Comet-Style Hadoop Interface

Within FIWARE, Comet is used to access historical data. However, Comet only works with MongoDB. Thus, develop a component exposing the same query interface as Comet but using data from Hadoop

Excel2CSV automation

Provide a mean to automatically convert the daily Excel report from FeedbackNow into a CSV file that allows for automatic import into the SCDP

Re-factor Service Publications

Internal, critical services like Orion, Comet, Cygnus, IDAS, ...) should only be accessible via PEP. Thus, make sure that these services are only reachable within nodes on Kubernetes (Cluster-IP)

Define initial set of data policies

Define a set of rules for all the data stored in the SDPG (FeedbackNow, Traffic, Air Quality) such that data can stored in an efficient way over a long period of time

Polling SOAP interface for traffic data

The SOAP Interface provided by the traffic department needs to be queried in regular intervals and the the resulting data forwarded to the SCDP.
This could possibly also be done using NodeRED

Optimize HDFS Access from Cygnus

Cugnus is currently using the Hadoop Namenode's REST Service (WebHDFS), which is apparently way to slow for high loads. This should be refactored, using a "native" Hadoop interface.

NodeRED Evaluation

Check whether NodeRED is a suitable platform for full-custom code needed throughout this project

Provide HA with automatic failover for Hadoop Cluster

The NameNode is a Single Point of Failure. In order to provide high availability a standby namenode is setup.
Steps that have to be performed:

  • Create new docker images
  • Configure Automatic Failover with Zookeeper
  • Test HA cluster
  • Upgrade the cluster on Kubernetes
  • Migrate data

Data transfer from Mongo to Hadoop

Since the backup-system only supports Mongo, we need a decent way to get this data:

  • back into the production system (mongodump, mongorestore)
  • into hadoop (e.g. by converting the exported text file right into the Hadoop csv format)

Optimize FIWARE Deployment

Different deployment layouts shall be developed and performance tested in order to find one with an optimal throughput

Inbound REST-Interface for Air-Quality Sensor Data / T-Matix

Data collected by the proprietary IoT System connected to the Air-Quality sensor network is pushed in real-time to our FIWARE IoT-Platform. This REST-Interface will receive this data, translate its payload into Ultralight and send it to IDAS.

Setup internal Node-RED instance for rewriting IDAS URLs

IDAS requires API-Key and device ID to be passed as query parameters. This disables XACML rules. Thus, Node-RED shall be used to rewrite URL-Parameters to Query-Parameters.
PEP will then speak to Node-RED rather than IDAS

Heartbeat for NodeRED Endpoints

For automatically testing the availability of our NodeRED end-points there should be special request that does not produce any data but returns 200 if everything is OK. This request should be sent by systems like Nagios

Install Hadoop on Kubernetes

A major part of the SDPG will be Hadoop and its relevant Plugins. All required nodes should be rolled out on the local cloud using kubernetes

Provide ML Samples

Setup some sample notebooks demonstrating the use of machine learning on SDPG data

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.