title

text

March 15 – 17 , 2017

Postrelease

  • more than
    0 participants
  • 0 speakers
  • 0
    minutes of conversation
  • 63 talks
  • offline
    format

Talks

Talks archive

PgConf.Russia 2017
  • Ivan Frolkov
    Ivan Frolkov PostgresPro

    With big data threads, even the upload of data to a database can often be problematic – apart from the data upload itself, you need to create indexes, perform VACUUM after the upload for correct Index-only scans, etc. From this talk, you will learn how to avoid most of such problems (if not all of them).

    VIDEO

  • Philip Delgyado
    Philip Delgyado ITIS Ltd

    When working with a complex business logic, you often have to implement a workflow - a sequence of several processing steps, with each step implementing a separate part of the business logic. This is usually done with specialized queues, but if there are high reliability demands, it makes sense to do everything on PostgreSQL.

    I will describe the tasks that require a workflow implementation, offer a solution, compare it with other options, and tell you about the implementation traps and pitfalls.

  • Egor Rogov
    Egor Rogov PostgresPro

    Modern state of backup tools for PostgreSQL certainly leaves some room for improvements. Standard utilities provides only basic functionality, third-party tools solve some, but not all, problems. To take backups reliably, fast, and incrementally on page level, one needs not only a backup tool, but also some support from the database server. On this tutorial we will talk about our new backup and recovery manager ProBackup, and will show it in action.

    Demonstration script:

  • Radoslav Glinsky
    Radoslav Glinsky Skype (Microsoft)

    Do you test your PostgreSQL releases prior to Production in a dedicated test environment? Are you sure that your test environment (shortly Test) is equal to Production and in an appropriate state?

    In Skype we were facing multiple challenges associated with database testing:
    - Simplifying complex Production architecture of thousands of PostgreSQL instances, interconnected with RPCs and replications, infrastructure servers and external DB scripts, into their Test counterparts.
    - Constantly growing hardware requirements, insufficient cleanup of data generated in Test.
    - Differences between Test and Production were appearing and accumulating. Recognizing and fixing them required lots of effort.

All talks