February was a month of very intensive work to prepare our straw detector and Data Acquisition System for tests with proton beam from COSY accelerator at Juelich Forschungszentrum in Germany.
Together with 5 other groups we had granted one week of beamtime to evaluate the detectors, electronics and software.
It was the first time we evaluated operation of the entire, small scale detector system for PANDA experiment. Three detector subsystems: Forward Tracker, Electromagnetic Calorimeter and Time-of-Flight, each with their own readout system, were synchronized with SODANet system and generated data was processed by a set of 3 Compute Node modules for burst building and preliminary preprocessing.
It was also the possibility to test the data preprocessing system based on Xilinx ZCU102 platform. The board receives data streams from the digitizing boards and recovers track candidates, rejecting empty events.
Additive Synthesizer – Gdansk University of Technology was granted with Digilent special prize for best usage of Digilent Instruments.
The event lasts for two days. On Saturday all finalists were presenting their work. On Sunday only few teams were invited for showing in details their solutions.
Most of submissions for the contest were diploma thesis or long-term projects conducted by experienced engineers.
Our approach for competition was to learn a lot and explore different approaches to Augmented Reality on SoC devices. In my opinion it is not necessarily important to win, but to compete, cooperate and learn state-of-the-art techniques and methods. We draw conclusions and got valuable feedback from community.
We have successfully managed to construct the first ever mini PANDA DAQ system!
Two subsystems: FT straws and EMC are working together, synchronized by SODANet and processed by Burst Building Network constructed out of 3 Compute Node modules.
First cosmics were collected and tracks reconstructed!
Enhanced image reconstruction, including 3D and TOF functionalities has been successfully implemented entirely in the FPGA!
In programmable logic, we are finding LOR candidates and reconstruct the annihilation point coordinates. Then, only X, Y, Z values are being sent from the JPET Controller to the server that produces 3D canvas with the scanner visualization.
You can find a video showing it in action under [this] link.