Stamane Giovanni Alessandroni ha discussio la prima tesi di laurea (in statistica) su Bologna “Città 30”.
--
#Bologna #Città30 #Statistics #BigData
Stamane Giovanni Alessandroni ha discussio la prima tesi di laurea (in statistica) su Bologna “Città 30”.
--
#Bologna #Città30 #Statistics #BigData
Twi-X missbraucht sensible Daten der Benutzer/innen
Mehrere NGO haben bei verschiedenen Aufsichtsbehörden Beschwerde eingelegt, weil Twi-X die DSGVO und den DSA verletze. Der verbietet es, Informationen über politische Einstellung oder Gesundheit einer Person für zielgerichtete Werbung
Announcing Python-Blosc2 3.6.1
!Unlock new levels of data manipulation with Blosc2!
We've introduced a major improvement: powerful fancy indexing and orthogonal indexing for Blosc2 arrays.
We've tamed the complexity of fancy indexing to make it intuitive, efficient, and consistent with NumPy's behavior.
Read all about it on our blog! https://www.blosc.org/posts/blosc2-fancy-indexing/
Compress Better, Compute Bigger!
Call for Papers!
Das vom @berliner_antike_kolleg mit DAI, @BBAW, @freieuniversitaet und @HumboldtUni organisierte #DigitalClassicistSeminar Berlin geht von Oktober '25 bis Februar '26 in die nächste Runde.
Insbesondere Nachwuchswissenschaftler:innen sind eingeladen, bis zum 16.08. Beiträge zu digitalen Methoden, Herausforderungen von #BigData, #Visualisierungen u.v.m. einzureichen.
Den vollständigen #CfP gibt es hier: https://digital-classicist.bbaw.de/cfp
What does it take to maintain one of the world's largest repositories of free, structured knowledge?
Read this interview about the challenges Wikidata faces and how the team handles massive scale and constant updates, all while remaining open source:
https://www.bigdatawire.com/2025/07/10/scaling-the-knowledge-graph-behind-wikipedia/ #OpenData #KnowledgeGraphs #BigData #SemanticWeb
Live Webinar am 31.07. zu Legacy-Modernisierung – mit LLMs praktisch umgesetzt
Erik Dörnenburg erläutert, wie man mit Reverse Engineering Tools in geschickter Kombination mit LLMs die Legacy Modernisierung beschleunigen kann.
@gerrymcgovern In the colonial world where #IndigenousPeoples have been pushed into #poverty, now the push is to take the #water and make marginalized people and #ecosystems complicit in their own demise. https://www.cbc.ca/news/canada/edmonton/first-indigenous-data-centre-abandoned-power-plant-1.7586072
Bisschen Gefluche zu #bigdata, #KI, Tech Bros, #hackerspace|s, warum wir uns Technik zurückerobern müssen und warum die #soziologie dazu forschen sollte.
Sollte 5min Science Slam werden, eskalierte, nun #blog
Blosc2 is incredibly powerful, but its flexibility comes with a challenge: navigating thousands of possible compression parameter combinations (codecs, levels, filters) to find the best one for your data.
Stop guessing and start optimizing. Save valuable time and unlock the full potential of your data compression pipeline with Btune
Compress Better
Blosc2 supports memory-mapped files for super-efficient data access!
Why memory-mapping?
No system call overhead for each read/write
Data goes straight from page cache to user space—much faster than traditional I/O!
https://github.com/Blosc/python-blosc2/blob/main/bench/io.py
Join our tutorial at
@EuroSciPy 2025, where we'll dive deep into these techniques and share more expert tips for maximizing data throughput. See you there!
#Allgemein #Empfehlung #Mobilfunk #Warnung #android #bigdata #chat #google #KI #messenger #privacy #privatsphäre #smartphone #spionage #UnplugTrump #vorbeugen #wissen #customrom
#GrapheneOS ist eine - bequeme - Lösung für dieses Problem
Fortschritt: Google Android spioniert jetzt noch brutaler.
Google informiert seine Nutzer/innen, dass sie ab sofort durch die KI Gemini unterstützt werden - ob sie es wollen oder nicht. Gemini erhält dafür Zugang zu allen Apps, auch zu verschlüsselnden Messengern. Lokal liegen die Inhalte ja im Klartext vor, und darauf will Gemini zugreifen. Ist das der Anfang von Quellen-TKÜ (vulgo Chatkontrolle)? Googles Vorstoß riecht ziemlich streng nach Salami-Taktik in diese Richtung.
@jomla Yes, definitely some #BigData vibes there. Same story, 10-15 years ago. We should learn from it. What we need instead of hype or doom, imho, is intense, critically-engaged, task-oriented evaluation to see what works, what is helpful, what are unhelpful scenarios, how this stuff transfers to our domain, etc. Many in #DH are starting to do just that, so I am hopeful.
@NatObserver @canada-s-national-observer-NatObserver #Canada's Corporate Conservative Party, bought and paid for by Trump, #BigOil, #BigTech #BigData, #BigBanks
Jak przetworzyć 10 milionów plików PDF z ministerialnych raportów i zbudować z tego działający serwis publiczny? Tomasz Zieliński opowiedział o kulisach powstania SLPS - od walki z danymi, przez ekstrakcję tekstu i analizę niezgodności, po budowę wydajnego systemu i jego niskokosztowy hosting.
Zobacz wykład z OMH 2024 i przygotuj się na tegoroczną edycję: https://www.youtube.com/watch?v=PJHNkX0XP50
Rejestracja na Oh My Hack 2025 trwa
https://bit.ly/OMHbilet
Announcing Python-Blosc2 3.5.1
We, Blosc developers, understand that memory efficiency is critical when working with large datasets. To that end, we continuously profile and optimize our codebase to deliver the best possible performance.
This version introduces significant performance and memory optimizations, enhancing the experience of computing with large, compressed datasets.
Compress Better, Compute Bigger!
Meta’s AI copyright win comes with a warning about fair use
“This ruling does not stand for the proposition that Meta’s use of copyrighted materials to train its language models is lawful,” Judge Chhabria said.
“It stands only for the proposition that these plaintiffs made the wrong arguments and failed to develop a record in support of the right one.”
#meta #LlamaAI #artificialintelligence #AI #LLM #data #bigdata #l3gal #copyright #fairuse #technology #tech
https://www.theverge.com/news/693437/meta-ai-copyright-win-fair-use-warning
Using dplyr and ggplot2 in R can significantly streamline your data analysis process, making it easier to work with complex data sets.
I have created a video tutorial in collaboration with Albert Rapp, where I demonstrate how to do this in practice: https://www.youtube.com/watch?v=EKISB0gnue4