10 interesting stories served every morning and every evening.
Watch: Cruise ships chopped in half are a license to print money
Watch: Cruise ships chopped in half are a license to print money
It’s massively profitable for cruise operators to hack their ships in half and stick an extra section in to lengthen them
It’s massively profitable for cruise operators to hack their ships in half and stick an extra section in to lengthen them
The new section slides in, complete with furniture and fittings that might otherwise be impossible to load on board
That extra section costs somewhere around US$80 million. It doesn’t look like much, but it radically transforms the ship’s money-making capability
A super-precise cut right through the middle of the ship
View gallery - 4 images
Slicing huge cruise ships in half, then welding in an extra segment to lengthen them, is more or less a license to print money for cruise operators — so this ‘jumboization’ surgery is becoming very common. Let’s take a look at how it’s done. Humanity, it seems, can’t get enough of cruise ships. In 1990, according to Cruise Market Watch, the global cruise industry treated some 3.7 million passengers to its familiar regime of buffet food, all-inclusive child supervision, shuffleboard, plentiful liquor and winking entertainers. In 2024, that number’s tracking closer to 30 million.So the ships are getting bigger, as evidenced by Royal Caribbean Group’s Icon of the Seas, launched this January. This gaudy, teetering behemoth is designed to carry just under 10,000 people, including crew. It’s just under 1,200 ft (365 m) long, 159 ft (48 m) wide, and it’s stacked no less than 20 decks high, with a technicolour water fun park, ‘Aquadome,’ seven swimming pools and the industry’s tallest waterfall piled on top. Good lord:But this towering leisure-fest cost around US$2 billion to build, and took around two and a half years to launch from the first steel cutting. The cruise ship industry can’t build new ships fast enough to satisfy the rampant demand — and building new ships also means you have to train new crews.
The cheaper, easier way for operators to expand their carrying capacity and jack up profits is to make existing ships bigger, as it turns out. For an average of around $80 million, and just a couple of months out of service, operators can chop an existing ship down the middle, slide in a new slice that’s designed to fit perfectly, weld it together, and come away with enough extra premium cabins to pay off the whole operation within a few years. That’s not to mention the opportunity for a new paint job, bigger deck pools or engine upgrades while the ship’s up on blocks — and HR only needs to train a small percentage of extra staff to add to an existing crew. The result: with a much smaller outlay and a negligible gap in service, operators can make an existing boat much more profitable. ‘Jumboization,’ as the process is known, certainly isn’t a new idea, or unique to cruise ships. Indeed, the term itself appears to have been coined just after World War II, at which point shipbuilders were already doing it at enormous scale to lengthen warships, in some of the most complex engineering work ever attempted in the slide rule era. Nowadays, it’s a well-trodden path, with certain shipyards specializing in the process of adding 80-130 ft (24-40 m) to a given cruise ship. The process takes the best part of a year if you include the measure, design and build of the new slice of ship — but remarkably, just a matter of weeks for the actual major surgery itself. How it’s doneFirst, you need the new chunk of ship ready to roll. Engineers pore over the initial design plans, and then go on board to make hundreds of measurements; clearly, it’s got to be precise.This done, they design the new section. Every one of the thousands of pipes, wires, cables and ventilation channels that’ll be cut are designed into the new section, ready to line up perfectly at both ends when the new joins are made. In this way, existing functionality can be maintained — and if the extra cabins place too much demand on existing systems, they can be upgraded to handle their new loads. Then, they go ahead and build the new section, often complete with interior fit-outs, at a shipyard facility, leaving the ends open. Sometimes, as in the video below of the MSC Armonia, these open sections are actually christened and launched into the water in their own right, and towed to the dry dock where the surgery is scheduled, like floating apartment blocks. Then, the ship is brought in, and precisely positioned above an array of “skid shoes” — little ship-lifting sleds positioned on metal tracks pre-bolted to the floor of the dry dock. Each of these little sleds is a hydraulic jack capable of lifting as much as 1,000 tons — and each is also capable of applying hydraulic lateral force to move it along the tracks. There might be upwards of 50 of these skid shoes involved in a given ship-stretching operation. With the ship positioned over the skid shoes, the dock is emptied, and the ship comes down to rest on its new mobile supports. Then, a work crew that could extend into the hundreds gets to work chopping the ship in half. Laser guides are used to ensure millimeter-level precision, since it’s critical that the edges meet perfectly when the new section comes in. And of course, care is taken to plan which bits get cut in which order, because of the enormous weight involved and the associated structural pressures that might result.
A super-precise cut right through the middle of the ship
The cutting job itself is split between automated cutting machines, which handle the straight surfaces like the side walls and floors, and skilled laborers for curved or complex shapes, but both the machines and the workers use common acetylene blowtorches to cut through the metal.When the whole ship’s been cut right through (memories of a particularly horrifying Three Body Problem scene anyone?) the skid shoes begin the co-ordinated process of pulling the two halves apart, far enough for the new slice to be moved in on its own team of skid shoes. This is also a fine opportunity to load the existing ship segments up with large fitments and items that would otherwise be impossible to get into the lower decks.The whole ship can be chopped in half and separated in about two days.
The new section slides in, complete with furniture and fittings that might otherwise be impossible to load on board
Once that’s all done, the edges of the stem, middle and stern sections are lined up to perfection using the lifting and moving capabilities of the skid shoes, and then it’s time to start connecting things back together. The surfaces are brought together, and the welding begins.A combination of electrode and stick welding is used — and here’s where the extreme cutting precision really comes into play, because the maximum distance that can be covered between the two surfaces is only a few millimeters. The closer, the better — and of course, there’s really no margin for error with the watertight hull and the structural supports that’ll hold the massive ship together in rough seas. With the hull, decks and structure all joined together, it’s then time to reconnect all the thousands of cables, ducts, wires and other fitments, while interior teams scurry to make the joins seamless on the inside of the ship, and another squad begins to give the whole ship a paint job that’ll cover up the welds and scars on the exterior. The shipyard’s responsibility is structural integrity and seaworthiness — but the operator also has to thoroughly test every single electrical, mechanical and hydraulic system on the ship to make sure everything has been reconnected properly. They also need to run sea trials to prove the strength and handling of the new, longer ship.The whole operation, from when the ship arrives at the dry dock to the point where it’s approved to be sent back out into service, can take just nine weeks. If you include the time spent building the new section, as well as all the pre-planning and engineering, the whole project might run out to about nine months.
That extra section costs somewhere around US$80 million. It doesn’t look like much, but it radically transforms the ship’s money-making capability
As I said, this isn’t a new process; Wikipedia lists 21 cruise ships that have been lengthened in this way since 1977, and indeed the mighty Seawise Giant — the longest self-propelled ship in history at a staggering 1,504 ft (458 m) end to end — owes some of its prodigious length to the jumboization technique.But the scale of these operations, the techniques involved and the financial incentives behind them have made it a fascinating rabbit hole for me. I hope you’ve enjoyed it too!And if you wanna dive even deeper, take a look at this fantastic mini-doco from Spark:
Jumboization - Modifying Ships To Make Them Even Bigger [4K] | Heavy Lift | Spark
View gallery - 4 images
Loz leads the New Atlas team as Editorial Director, after nearly two decades as one of our most versatile writers. He’s also proven himself as a photographer, videographer, presenter, producer and podcast engineer. A graduate in Psychology, former business analyst and touring musician, he’s covered just about everything for New Atlas, concentrating lately on clean energy, AI, humanoid robotics, next-gen aircraft, and the odd bit of music, motorcycles and automotive.
...
Read the original on newatlas.com »
Old codger yells at software, part the latest.
The story so far
A great amount of my writing on this site revolves around complaining about modern user interface design. Some people agree with me, some don’t. Most probably don’t care at all. That’s fine. What I find interesting is that many (but not all!) people who disagree either present extremely specific non-argument nitpicks like “Something in Windows 3.1 was bad, too!” and ignores the big questions posed, such as whether it’s actually a good idea shoving everything into a single hamburger menu, or if mixing touch and desktop paradigms wildly between and sometimes within programs is actually beneficial to end users. Others - despite my efforts to the contrary - use sweeping dismissals of the kind “you just don’t like flat design.” That’s true - I don’t like flat design, but many of my arguments have nothing whatsoever to do with aesthetics.
Well, that’s what writing on the net is like. But I shall not despair, nor shall I be silenced! Allow me, for a few moments, to focus on a very detailed example that’s got nothing to do with flatness, but rather with how to access core program functionality.
It’s worth mentioning that I agree that the modern design paradigm probably is friendly to beginner users in many ways. But at some point, people stop being beginners. People who use computers several hours per day, performing a wide variety of tasks in many different programs, should also be taken in to account when designing software. As such, my critique comes from the point of what’s usually called a “power user”. It’s also worth considering that the more an interface hides, the less it offers by way of opportunities for a user to grow and learn.
For full transparency: It’s no secret that I don’t particularly enjoy the Gnome desktop. When I installed Ubuntu on my work laptop about a year ago, I wanted to see how long I could get by using Gnome. The answer was “about five minutes”. When I decided to switch the default desktop background image to a solid color, I discovered that the only available option in the configuration tool was to select images. Sure, I could google a solution to the problem, which involved a pretty lengthy command line incantation, but depriving the user of a simple option to select a background color was more than my patience could handle.
I’ve written some pretty harsh critique about Gnome before, but I feel this is a project that does deserve critique. It’s not just the default desktop environment in most major Linux distributions, it’s also a project that’s very vocal - opinionated, as the saying goes - about how to do things. A few examples might be in order:
“Our software is built to be usable by everyone. We care deeply about user experience.”
“Software should be structurally and aesthetically elegant.”
“People’s attention is precious. We pride ourselves in being distraction free.”
“The ‘traditional desktop’ is dead, and it’s not coming back. (…) Instead of trying to bring back old concepts like menu bars or status icons, invent something better from first principles.”
Bold words, indeed! Let’s see how they fare in the real world. Recently, in another sudden fit of open-mindedness, I decided to perform some light file organizing using Gnome Files. This isn’t just any old Gnome application: the file manager is a central point of any desktop environment, along with window management and a decent set of UI widgets. Surely the Gnome project will have spent a good amount of time and effort on making Gnome Files a flagship application, showcasing their superior UI philosophy from its absolutely best side.
Well, it certainly doesn’t look too shabby. Clean, elegant, even minimalist. A calming appearance. Sure, it’s a flat design, but I can clearly make out the different UI elements and get a good idea of what’s clickable or not - or so I’m led to believe. There’s just one problem - I’d prefer a list view instead of the large icons. How do I change it?
Note the many distinct-looking icons in the toolbar. From left to right we have three stacked dots, three stacked dots and lines, and three stacked lines. Pretty obvious what they’re all about, right? Anyways, this dropdown menu with the tooltip text “View Options” is probably a good start.
Alas, that menu only shows various sort options. This makes me think “Sort Options” would have been a better tooltip text. Then again, I’m not much for reinventing everything from first principles, so what do I know?
What’s in the “Main Menu”, then? Nothing that has to do with a list view, I’m afraid - though there are options here I’d expect to find in a menu called “View Options”, such as Icon Size and Show Hidden Files.
There’s a “Folder Menu” too, but that doesn’t help me much, either.
Ah! There it is! The “View Options” dropdown is actually a split button, with a toggle part and a dropdown part. That’s sort of fine I guess, but I’d expect the toggle options to also be listed in the dropdown portion of the widget. If not, why group them together like this? Why not split them into two clearly separate buttons? This honestly took me a good while to figure out - and I’d like to think I’m neither extremely stupid nor particularly inexperienced with computers.
Yes, it really was this hard for me to find the list view toggle. The “View Options” dropdown didn’t contain view options, but rather sort options, and I didn’t realize it was a split button with two completely different functions. I was further confused when other actual view options were, in fact, only available in the “Main Menu”. To me, this doesn’t feel like “structural elegance”. By now, my calm was replaced with frustration.
I actually resorted to the built-in help function when looking for a way to switch to the list view. I searched for “list view”, but didn’t find anything immediately helpful. Even though I brought up the help window using Gnome Files’ menu option for this, search results were listed for many other applications. Search result number nine, “Browse files and folders”, was at least related to Gnome Files - but appeared after things like “Manage volumes and partitions” and “Edit contact details”. The only entry I could find while manually browsing the help that mentioned “List View” was about what I could do when the list view had already been selected. I tried my best, but couldn’t find any information on how to actually activate it.
While browsing the help, I noticed something else: frivolous tooltips. Tooltips are a great invention and can be very useful, but a lot of programs nowadays seem to just sprinkle them everywhere, with no particular thought as to whether they’re actually meaningful or not. Gnome Help is one such case.
Here, I’m just resting my mouse pointer somewhere, as one might do when reading through a bunch of clickable items. Up pops a tooltip that not only completely covers the title of the next item, it also has the exact same text as the header of the item it belongs to. Why? Perhaps it’s part of the “distraction free experience”.
The same goes for the left hand side toolbar in Gnome Files. Yes, I assumed that “Recent” and “Starred” here, in the context of Gnome Files, are about recent and starred files and not something else. Granted, some of these tooltips show a full path, but honestly - if I’ve added something to this bar, I probably know what it is and where it’s located. A tooltip would be useful if two items have the same name - but why show them everywhere else? To me, at least, it’s very annoying that I can’t rest my mouse pointer somewhere without having pointless information popping up that covers what I’m really interested in looking at. In fact, I’d say this behavior teaches me as a user that tooltips are pointless and irritating, making me instinctively ignore them even when they might be useful.
Navigating using the Gnome Files UI is fine, for the most part. I’m unused to the button locations, and I miss a button for going one level up, to the parent directory. There are buttons for going back and forward in the navigation history, but that’s not the same thing. Clicking on a directory in the location bar will take you there, but it’s much easier to misclick and thus less convenient.
When searching for a “parent directory” button, I noticed a few strange behaviors. Clicking on directory names in the location bar is a nice feature. However, many file managers will also let the user directly enter a path here. In fact, it looks a lot like a text box (such as the Folder name widget in the screenshot above) but no matter where I click in it, I cannot activate a normal editing mode.
It seems this editing mode can only be activated using a keyboard shortcut, Ctrl-L, which isn’t immediately apparent - or, to be frank, very logical. It looks like a text box, and it is indeed a text box - not just all the time. Maybe I’m not supposed to think of it as a text box, but the design language here isn’t exactly super clear on what kind of input I’m dealing with, so I’m working from assumptions based on having used computers and GUI file managers for 36 years - which is something developers and designers should take into account when building interfaces.
Having no mouse-driven way of activating a fairly common UI task feels like a step back in discoverability. In fact, I thought the feature hadn’t been implemented until I googled it. To be fair, it listed in the Keyboard Shortcuts window, but call me old school for expecting GUI elements to be accessible with the mouse.
The shortcuts list is three pages long. It’s got a search function, which is nice - if you know what you’re looking for. For example, Gnome Help calls the location bar “the path bar”, but searching for “path” gives zero results. If you don’t know what to search for, there’s no table of contents or other way of quickly viewing the various categories of shortcuts. You have to examine each of the pages and hope to find what you’re looking for.
Upon reaching page three, I learned that there’s even a keyboard shortcut for opening the keyboard shortcuts window - but unlike some other shortcuts, it’s not shown next to its corresponding GUI entry in the Main Menu. Inconsistent and confusing.
I’m not sure why, but the Gnome project has a strong aversion to menu bars. The keyboard shortcuts window may be “reinventing from first principles”, but to me this invention looks like a flawed reinterpretation of a traditional menu bar. A menu bar provides a way of exposing categorized program features in a familiar and always-visible UI, lists their keyboard shortcuts in a consistent manner and it lets the user interact with the option immediately upon finding it.
In Gnome Files, we’re instead given a handful of features scattered across the UI. Hidden features (accessible solely through keyboard shorcuts) can only be learned by browsing what is best described as a non-interactive menu of the kind you’d find printed on paper in a restaurant. This browsing brings a considerable context switch because the window is modal, so you can’t keep it open while experimenting. You have to find what you’re looking for, note the keyboard shortcut, close the shortcuts window, and then invoke the feature. Keyboard shortcuts aren’t bad, but in a mouse driven environment, having no way of finding and invoking features through the GUI is limiting and confusing. And why force users to memorize a shortcut for something they might only do occasionally?
In short, this feels like an afterthought rather than a revolutionary new, efficient approach to operating a GUI.
When trying to activate path editing using the mouse, my vigorous clicking just made the Gnome Files window move around. Since the window has no real title bar, it must be moved by clicking-and-dragging in the top part of the window. This is also where the toolbar resides, which means click-dragging on UI controls that already have a completely different function. You can use the search function by clicking on the search icon - but you can also click the search icon and start dragging the window around.
There’s a location history available by either context-clicking or performing a “long click” on the back- and forward buttons, but nothing about the buttons indicate this (and the feature doesn’t seem to have a keyboard shortcut), which introduces further ambiguity regarding the click-drag behavior. Confusingly, performing a long click on other items doesn’t bring up their context menu.
Activating windows with the mouse is even worse: If I want to bring a window to the front of the stack, I have to search for a non-clickable area, lest I activate the search feature, skip upwards in the path, switch from list view back to icon view or accidentally access some other program feature. This introduces extreme ambiguity that makes me as a user feel insecure and uncomfortable: a simple mouse slip can give the most surprising results. The user is conditioned to be mindful of this, which increases the cognitive load for very simple and common window operations.
Another curious idiosyncrasy of forgoing window title bars is that right-clicking on for example the search icon actually brings up a window management menu, with options for closing, maximizing and minimizing the window. This pops up when right clicking anywhere in the top part of the window. Except when right clicking on a directory name in the location bar - that will instead bring up a completely different menu, with various directory actions. Except if you click on the current directory, which won’t bring up a context menu. Except sometimes it does, but then it’s the window actions menu. Which is also what happens if you miss the context-clickable area of another directory by just a pixel, even if the pointer is still inside the location bar.
You can, however, middle click on any directory name - - to open it in a new tab. Which is also one of the options available in the context menu.
Gnome Files, or rather GTK 4, uses hidden scroll bars. I’m using what I assume to be the default settings and the default GTK 4 theme as supplied by Debian, because I haven’t changed anything (except the font). I personally don’t like hidden scroll bars, because hiding the scroll bar also hides information not only about what I can do with the GUI itself, but also about where I’m currently positioned in E. G. a file listing or text document.
Moving the mouse about in Gnome Files reveals the scroll bar, as depicted in the first picture above. It’s very small and very hard to see because of the low contrast, but it’s there. But - Ah-hah! - the scroll bar is actually hidden in two steps! When moving the mouse pointer to it, it grows and becomes more visible, as in the second picture above. But when it does, it also moves the entire width of its original size to the left, meaning that my mouse pointer is now pointing at… nothing. Thanks, Gnomebama.
The Gnome Files UI feels haphazard, incoherent and sometimes even dangerous:
Menu names and their contents are confusing, with “View Options” actually being sort options. Actual view options are instead sprinkled across other parts of the UI.
Some common features are only accessible - and discoverable - through keyboard shortcuts. The keyboard shortcuts listing is non-interactive, modal, and incurs a substantial mental context switch.
Widgets don’t behave quite as you’d expect when compared to other, similar, modern applications.
Even within the closed world of Gnome Files, widget looks can be deceptive.
Tooltips are either misleading, or comically uninformative and thus annoyingly distracting.
Moving windows by clicking on icons that already have a specific function feels unintuitive and introduces an unnecessary risk of misclicking.
To activate a window without also activating a program feature, time must be spent searching for a non-interactive area of the UI.
Context-clicking in the top part of the window gives spurious and unpredictable results.
In the default theme, scroll bars jump away from their original position when you point at them.
Searching and then browsing the built-in help for “list view” didn’t actually help me find out how to enable the list view. The only page I could find that contains the exact phrase appears as #15 among the search results.
Nomenclature differs between the help texts and the actual GUI.
The many various inconsistencies listed above makes it hard for me as a user to construct mental models and patterns for making efficient and reliable assumptions about the UI. I often have to double check things and look for features in the keyboard shortcuts window, since I don’t know if they’ve been deigned a menu entry or toolbar icon. Ambiguous mouse behavior introduces unnecessary cognitive load, making it hard to build confidence even when performing basic tasks.
Designing user interfaces is hard. Writing software is hard. Designing great interfaces and writing great software is even harder. No matter how hard you try, you can’t please everyone. Functionally, Gnome Files lets me do file management. I could, if I had no other options, probably get used to its UI idiosyncrasies. But as I’ve hopefully demonstrated above, there are many things about Gnome Files - a central part of the Gnome desktop - that can be considered objectively bad from a UI design standpoint.
This doesn’t mean that there aren’t other bad programs or that previous design paradigms were perfection incarnate. Neither does it mean that every little bit of Gnome Files is worthless, or that Gnome and Gnome Files are unique in employing these patters (but that’s hardly a comfort). But shouldn’t this new design paradigm produce something better? Isn’t that its whole raison d’être? After ten-fifteen years of the old mainstream desktop paradigm, we arrived at Windows 95 - which wasn’t perfect, but pretty damn consistent, predictable and reliable. After ten-fifteen years of this new paradigm, we instead have constant redesigns and reworkings of solutions to problems we already know how to solve.
Examining these new solutions suggests to me that maybe proponents of the new UI paradigm should be a bit more humble in their approach to tried and tested patterns. Nearly all of the criticisms examined above already have well known solutions that have been honed for a period of decades. For example: Having actual window title bars, consistently listing keyboard shortcuts in menus, consistently categorizing menus and options, and using a richer design language. Old doesn’t automatically mean worse, just as new doesn’t automatically mean better.
My personal conclusion is that if this is the result of inventing “something better from first principles” to create “structurally elegant” and “distraction free” software “usable by everyone”, I guess I’m just not “everyone”. And that would be fine, if it wasn’t for the fact that these days, I’m often left with no choice but using programs with UI:s like these.
...
Read the original on www.datagubbe.se »
Academic journals, archives, and repositories are seeing an increasing number of questionable research papers clearly produced using generative AI. They are often created with widely available, general-purpose AI applications, most likely ChatGPT, and mimic scientific writing. Google Scholar easily locates and lists these questionable papers alongside reputable, quality-controlled research. Our analysis of a selection of questionable GPT-fabricated scientific papers found in Google Scholar shows that many are about applied, often controversial topics susceptible to disinformation: the environment, health, and computing. The resulting enhanced potential for malicious manipulation of society’s evidence base, particularly in politically divisive domains, is a growing concern.
Where are questionable publications produced with generative pre-trained transformers (GPTs) that can be found via Google Scholar published or deposited?
What are the main characteristics of these publications in relation to predominant subject categories?
How are these publications spread in the research infrastructure for scholarly communication?
How is the role of the scholarly communication infrastructure challenged in maintaining public trust in science and evidence through inappropriate use of generative AI?
A sample of scientific papers with signs of GPT-use found on Google Scholar was retrieved, downloaded, and analyzed using a combination of qualitative coding and descriptive statistics. All papers contained at least one of two common phrases returned by conversational agents that use large language models (LLM) like OpenAI’s ChatGPT. Google Search was then used to determine the extent to which copies of questionable, GPT-fabricated papers were available in various repositories, archives, citation databases, and social media platforms.
Roughly two-thirds of the retrieved papers were found to have been produced, at least in part, through undisclosed, potentially deceptive use of GPT. The majority (57%) of these questionable papers dealt with policy-relevant subjects (i.e., environment, health, computing), susceptible to influence operations. Most were available in several copies on different domains (e.g., social media, archives, and repositories).
Two main risks arise from the increasingly common use of GPT to (mass-)produce fake, scientific publications. First, the abundance of fabricated “studies” seeping into all areas of the research infrastructure threatens to overwhelm the scholarly communication system and jeopardize the integrity of the scientific record. A second risk lies in the increased possibility that convincingly scientific-looking content was in fact deceitfully created with AI tools and is also optimized to be retrieved by publicly available academic search engines, particularly Google Scholar. However small, this possibility and awareness of it risks undermining the basis for trust in scientific knowledge and poses serious societal risks.
The use of ChatGPT to generate text for academic papers has raised concerns about research integrity. Discussion of this phenomenon is ongoing in editorials, commentaries, opinion pieces, and on social media (Bom, 2023; Stokel-Walker, 2024; Thorp, 2023). There are now several lists of papers suspected of GPT misuse, and new papers are constantly being added. While many legitimate uses of GPT for research and academic writing exist (Huang & Tan, 2023; Kitamura, 2023; Lund et al., 2023), its undeclared use—beyond proofreading—has potentially far-reaching implications for both science and society, but especially for their relationship. It, therefore, seems important to extend the discussion to one of the most accessible and well-known intermediaries between science, but also certain types of misinformation, and the public, namely Google Scholar, also in response to the legitimate concerns that the discussion of generative AI and misinformation needs to be more nuanced and empirically substantiated (Simon et al., 2023).
Google Scholar, https://scholar.google.com, is an easy-to-use academic search engine. It is available for free, and its index is extensive (Gusenbauer & Haddaway, 2020). It is also often touted as a credible source for academic literature and even recommended in library guides, by media and information literacy initiatives, and fact checkers (Tripodi et al., 2023). However, Google Scholar lacks the transparency and adherence to standards that usually characterize citation databases. Instead, Google Scholar uses automated crawlers, like Google’s web search engine (Martín-Martín et al., 2021), and the inclusion criteria are based on primarily technical standards, allowing any individual author—with or without scientific affiliation—to upload papers to be indexed (Google Scholar Help, n.d.). It has been shown that Google Scholar is susceptible to manipulation through citation exploits (Antkare, 2020) and by providing access to fake scientific papers (Dadkhah et al., 2017). A large part of Google Scholar’s index consists of publications from established scientific journals or other forms of quality-controlled, scholarly literature. However, the index also contains a large amount of gray literature, including student papers, working papers, reports, preprint servers, and academic networking sites, as well as material from so-called “questionable” academic journals, including paper mills. The search interface does not offer the possibility to filter the results meaningfully by material type, publication status, or form of quality control, such as limiting the search to peer-reviewed material.
To understand the occurrence of ChatGPT (co-)authored work in Google Scholar’s index, we scraped it for publications, including one of two common ChatGPT responses (see Appendix A) that we encountered on social media and in media reports (DeGeurin, 2024). The results of our descriptive statistical analyses showed that around 62% did not declare the use of GPTs. Most of these GPT-fabricated papers were found in non-indexed journals and working papers, but some cases included research published in mainstream scientific journals and conference proceedings. More than half (57%) of these GPT-fabricated papers concerned policy-relevant subject areas susceptible to influence operations. To avoid increasing the visibility of these publications, we abstained from referencing them in this research note. However, we have made the data available in the Harvard Dataverse repository.
The publications were related to three issue areas—health (14.5%), environment (19.5%) and computing (23%)—with key terms such “healthcare,” “COVID-19,” or “infection”for health-related papers, and “analysis,” “sustainable,” and “global” for environment-related papers. In several cases, the papers had titles that strung together general keywords and buzzwords, thus alluding to very broad and current research. These terms included “biology,” “telehealth,” “climate policy,” “diversity,” and “disrupting,” to name just a few. While the study’s scope and design did not include a detailed analysis of which parts of the articles included fabricated text, our dataset did contain the surrounding sentences for each occurrence of the suspicious phrases that formed the basis for our search and subsequent selection. Based on that, we can say that the phrases occurred in most sections typically found in scientific publications, including the literature review, methods, conceptual and theoretical frameworks, background, motivation or societal relevance, and even discussion. This was confirmed during the joint coding, where we read and discussed all articles. It became clear that not just the text related to the telltale phrases was created by GPT, but that almost all articles in our sample of questionable articles likely contained traces of GPT-fabricated text everywhere.
Generative pre-trained transformers (GPTs) can be used to produce texts that mimic scientific writing. These texts, when made available online—as we demonstrate—leak into the databases of academic search engines and other parts of the research infrastructure for scholarly communication. This development exacerbates problems that were already present with less sophisticated text generators (Antkare, 2020; Cabanac & Labbé, 2021). Yet, the public release of ChatGPT in 2022, together with the way Google Scholar works, has increased the likelihood of lay people (e.g., media, politicians, patients, students) coming across questionable (or even entirely GPT-fabricated) papers and other problematic research findings. Previous research has emphasized that the ability to determine the value and status of scientific publications for lay people is at stake when misleading articles are passed off as reputable (Haider & Åström, 2017) and that systematic literature reviews risk being compromised (Dadkhah et al., 2017). It has also been highlighted that Google Scholar, in particular, can be and has been exploited for manipulating the evidence base for politically charged issues and to fuel conspiracy narratives (Tripodi et al., 2023). Both concerns are likely to be magnified in the future, increasing the risk of what we suggest calling evidence hacking—the strategic and coordinated malicious manipulation of society’s evidence base.
The authority of quality-controlled research as evidence to support legislation, policy, politics, and other forms of decision-making is undermined by the presence of undeclared GPT-fabricated content in publications professing to be scientific. Due to the large number of archives, repositories, mirror sites, and shadow libraries to which they spread, there is a clear risk that GPT-fabricated, questionable papers will reach audiences even after a possible retraction. There are considerable technical difficulties involved in identifying and tracing computer-fabricated papers (Cabanac & Labbé, 2021; Dadkhah et al., 2023; Jones, 2024), not to mention preventing and curbing their spread and uptake.
However, as the rise of the so-called anti-vaxx movement during the COVID-19 pandemic and the ongoing obstruction and denial of climate change show, retracting erroneous publications often fuels conspiracies and increases the following of these movements rather than stopping them. To illustrate this mechanism, climate deniers frequently question established scientific consensus by pointing to other, supposedly scientific, studies that support their claims. Usually, these are poorly executed, not peer-reviewed, based on obsolete data, or even fraudulent (Dunlap & Brulle, 2020). A similar strategy is successful in the alternative epistemic world of the global anti-vaccination movement (Carrion, 2018) and the persistence of flawed and questionable publications in the scientific record already poses significant problems for health research, policy, and lawmakers, and thus for society as a whole (Littell et al., 2024). Considering that a person’s support for “doing your own research” is associated with increased mistrust in scientific institutions (Chinn & Hasell, 2023), it will be of utmost importance to anticipate and consider such backfiring effects already when designing a technical solution, when suggesting industry or legal regulation, and in the planning of educational measures.
Solutions should be based on simultaneous considerations of technical, educational, and regulatory approaches, as well as incentives, including social ones, across the entire research infrastructure. Paying attention to how these approaches and incentives relate to each other can help identify points and mechanisms for disruption. Recognizing fraudulent academic papers must happen alongside understanding how they reach their audiences and what reasons there might be for some of these papers successfully “sticking around.” A possible way to mitigate some of the risks associated with GPT-fabricated scholarly texts finding their way into academic search engine results would be to provide filtering options for facets such as indexed journals, gray literature, peer-review, and similar on the interface of publicly available academic search engines. Furthermore, evaluation tools for indexed journals could be integrated into the graphical user interfaces and the crawlers of these academic search engines. To enable accountability, it is important that the index (database) of such a search engine is populated according to criteria that are transparent, open to scrutiny, and appropriate to the workings of science and other forms of academic research. Moreover, considering that Google Scholar has no real competitor, there is a strong case for establishing a freely accessible, non-specialized academic search engine that is not run for commercial reasons but for reasons of public interest. Such measures, together with educational initiatives aimed particularly at policymakers, science communicators, journalists, and other media workers, will be crucial to reducing the possibilities for and effects of malicious manipulation or evidence hacking. It is important not to present this as a technical problem that exists only because of AI text generators but to relate it to the wider concerns in which it is embedded. These range from a largely dysfunctional scholarly publishing system (Haider & Åström, 2017) and academia’s “publish or perish” paradigm to Google’s near-monopoly and ideological battles over the control of information and ultimately knowledge. Any intervention is likely to have systemic effects; these effects need to be considered and assessed in advance and, ideally, followed up on.
Our study focused on a selection of papers that were easily recognizable as fraudulent. We used this relatively small sample as a magnifying glass to examine, delineate, and understand a problem that goes beyond the scope of the sample itself, which however points towards larger concerns that require further investigation. The work of ongoing whistleblowing initiatives, recent media reports of journal closures (Subbaraman, 2024), or GPT-related changes in word use and writing style (Cabanac et al., 2021; Stokel-Walker, 2024) suggest that we only see the tip of the iceberg. There are already more sophisticated cases (Dadkhah et al., 2023) as well as cases involving fabricated images (Gu et al., 2022). Our analysis shows that questionable and potentially manipulative GPT-fabricated papers permeate the research infrastructure and are likely to become a widespread phenomenon. Our findings underline that the risk of fake scientific papers being used to maliciously manipulate evidence (see Dadkhah et al., 2017) must be taken seriously. Manipulation may involve undeclared automatic summaries of texts, inclusion in literature reviews, explicit scientific claims, or the concealment of errors in studies so that they are difficult to detect in peer review. However, the mere possibility of these things happening is a significant risk in its own right that can be strategically exploited and will have ramifications for trust in and perception of science. Society’s methods of evaluating sources and the foundations of media and information literacy are under threat and public trust in science is at risk of further erosion, with far-reaching consequences for society in dealing with information disorders. To address this multifaceted problem, we first need to understand why it exists and proliferates.
Finding 1: 139 GPT-fabricated, questionable papers were found and listed as regular results on the Google Scholar results page. Non-indexed journals dominate.
Most questionable papers we found were in non-indexed journals or were working papers, but we did also find some in established journals, publications, conferences, and repositories. We found a total of 139 papers with a suspected deceptive use of ChatGPT or similar LLM applications (see Table 1). Out of these, 19 were in indexed journals, 89 were in non-indexed journals, 19 were student papers found in university databases, and 12 were working papers (mostly in preprint databases). Table 1 divides these papers into categories. Health and environment papers made up around 34% (47) of the sample. Of these, 66% were present in non-indexed journals.
Finding 2: GPT-fabricated, questionable papers are disseminated online, permeating the research infrastructure for scholarly communication, often in multiple copies. Applied topics with practical implications dominate.
The 20 papers concerning health-related issues are distributed across 20 unique domains, accounting for 46 URLs. The 27 papers dealing with environmental issues can be found across 26 unique domains, accounting for 56 URLs. Most of the identified papers exist in multiple copies and have already spread to several archives, repositories, and social media. It would be difficult, or impossible, to remove them from the scientific record.
As apparent from Table 2, GPT-fabricated, questionable papers are seeping into most parts of the online research infrastructure for scholarly communication. Platforms on which identified papers have appeared include ResearchGate, ORCiD, Journal of Population Therapeutics and Clinical Pharmacology (JPTCP), Easychair, Frontiers, the Institute of Electrical and Electronics Engineer (IEEE), and X/Twitter. Thus, even if they are retracted from their original source, it will prove very difficult to track, remove, or even just mark them up on other platforms. Moreover, unless regulated, Google Scholar will enable their continued and most likely unlabeled discoverability.
A word rain visualization (Centre for Digital Humanities Uppsala, 2023), which combines word prominences through TF-IDF scores with semantic similarity of the full texts of our sample of GPT-generated articles that fall into the “Environment” and “Health” categories, reflects the two categories in question. However, as can be seen in Figure 1, it also reveals overlap and sub-areas. The y-axis shows word prominences through word positions and font sizes, while the x-axis indicates semantic similarity. In addition to a certain amount of overlap, this reveals sub-areas, which are best described as two distinct events within the word rain. The event on the left bundles terms related to the development and management of health and healthcare with “challenges,” “impact,” and “potential of artificial intelligence”emerging as semantically related terms. Terms related to research infrastructures, environmental, epistemic, and technological concepts are arranged further down in the same event (e.g., “system,” “climate,” “understanding,” “knowledge,” “learning,” “education,” “sustainable”). A second distinct event further to the right bundles terms associated with fish farming and aquatic medicinal plants, highlighting the presence of an aquaculture cluster. Here, the prominence of groups of terms such as “used,” “model,” “-based,” and “traditional” suggests the presence of applied research on these topics. The two events making up the word rain visualization, are linked by a less dominant but overlapping cluster of terms related to “energy” and “water.”
The bar chart of the terms in the paper subset (see Figure 2) complements the word rain visualization by depicting the most prominent terms in the full texts along the y-axis. Here, word prominences across health and environment papers are arranged descendingly, where values outside parentheses are TF-IDF values (relative frequencies) and values inside parentheses are raw term frequencies (absolute frequencies).
Finding 3: Google Scholar presents results from quality-controlled and non-controlled citation databases on the same interface, providing unfiltered access to GPT-fabricated questionable papers.
Google Scholar’s central position in the publicly accessible scholarly communication infrastructure, as well as its lack of standards, transparency, and accountability in terms of inclusion criteria, has potentially serious implications for public trust in science. This is likely to exacerbate the already-known potential to exploit Google Scholar for evidence hacking (Tripodi et al., 2023) and will have implications for any attempts to retract or remove fraudulent papers from their original publication venues. Any solution must consider the entirety of the research infrastructure for scholarly communication and the interplay of different actors, interests, and incentives.
We searched and scraped Google Scholar using the Python library Scholarly (Cholewiak et al., 2023) for papers that included specific phrases known to be common responses from ChatGPT and similar applications with the same underlying model (GPT3.5 or GPT4): “as of my last knowledge update” and/or “I don’t have access to real-time data” (see Appendix A). This facilitated the identification of papers that likely used generative AI to produce text, resulting in 227 retrieved papers. The papers’ bibliographic information was automatically added to a spreadsheet and downloaded into Zotero.
We employed multiple coding (Barbour, 2001) to classify the papers based on their content. First, we jointly assessed whether the paper was suspected of fraudulent use of ChatGPT (or similar) based on how the text was integrated into the papers and whether the paper was presented as original research output or the AI tool’s role was acknowledged. Second, in analyzing the content of the papers, we continued the multiple coding by classifying the fraudulent papers into four categories identified during an initial round of analysis—health, environment, computing, and others—and then determining which subjects were most affected by this issue (see Table 1). Out of the 227 retrieved papers, 88 papers were written with legitimate and/or declared use of GPTs (i.e., false positives, which were excluded from further analysis), and 139 papers were written with undeclared and/or fraudulent use (i.e., true positives, which were included in further analysis). The multiple coding was conducted jointly by all authors of the present article, who collaboratively coded and cross-checked each other’s interpretation of the data simultaneously in a shared spreadsheet file. This was done to single out coding discrepancies and settle coding disagreements, which in turn ensured methodological thoroughness and analytical consensus (see Barbour, 2001). Redoing the category coding later based on our established coding schedule, we achieved an intercoder reliability (Cohen’s kappa) of 0.806 after eradicating obvious differences.
The ranking algorithm of Google Scholar prioritizes highly cited and older publications (Martín-Martín et al., 2016). Therefore, the position of the articles on the search engine results pages was not particularly informative, considering the relatively small number of results in combination with the recency of the publications. Only the query “as of my last knowledge update” had more than two search engine result pages. On those, questionable articles with undeclared use of GPTs were evenly distributed across all result pages (min: 4, max: 9, mode: 8), with the proportion of undeclared use being slightly higher on average on later search result pages.
To understand how the papers making fraudulent use of generative AI were disseminated online, we programmatically searched for the paper titles (with exact string matching) in Google Search from our local IP address (see Appendix B) using the googlesearch–python library(Vikramaditya, 2020). We manually verified each search result to filter out false positives—results that were not related to the paper—and then compiled the most prominent URLs by field. This enabled the identification of other platforms through which the papers had been spread. We did not, however, investigate whether copies had spread into SciHub or other shadow libraries, or if they were referenced in Wikipedia.
We used descriptive statistics to count the prevalence of the number of GPT-fabricated papers across topics and venues and top domains by subject. The pandas software library for the Python programming language (The pandas development team, 2024) was used for this part of the analysis. Based on the multiple coding, paper occurrences were counted in relation to their categories, divided into indexed journals, non-indexed journals, student papers, and working papers. The schemes, subdomains, and subdirectories of the URL strings were filtered out while top-level domains and second-level domains were kept, which led to normalizing domain names. This, in turn, allowed the counting of domain frequencies in the environment and health categories. To distinguish word prominences and meanings in the environment and health-related GPT-fabricated questionable papers, a semantically-aware word cloud visualization was produced through the use of a word rain (Centre for Digital Humanities Uppsala, 2023) for full-text versions of the papers. Font size and y-axis positions indicate word prominences through TF-IDF scores for the environment and health papers (also visualized in a separate bar chart with raw term frequencies in parentheses), and words are positioned along the x-axis to reflect semantic similarity (Skeppstedt et al., 2024), with an English Word2vec skip gram model space (Fares et al., 2017). An English stop word list was used, along with a manually produced list including terms such as “https,” “volume,” or “years.”
Antkare, I. (2020). Ike Antkare, his publications, and those of his disciples. In M. Biagioli & A. Lippman (Eds.), Gaming the metrics (pp. 177–200). The MIT Press. https://doi.org/10.7551/mitpress/11087.003.0018
Barbour, R. S. (2001). Checklists for improving rigour in qualitative research: A case of the tail wagging the dog? BMJ, 322(7294), 1115–1117. https://doi.org/10.1136/bmj.322.7294.1115
Bom, H.-S. H. (2023). Exploring the opportunities and challenges of ChatGPT in academic writing: A roundtable discussion. Nuclear Medicine and Molecular Imaging, 57(4), 165–167. https://doi.org/10.1007/s13139-023-00809-2
Cabanac, G., & Labbé, C. (2021). Prevalence of nonsensical algorithmically generated papers in the scientific literature. Journal of the Association for Information Science and Technology, 72(12), 1461–1476. https://doi.org/10.1002/asi.24495
Cabanac, G., Labbé, C., & Magazinov, A. (2021). Tortured phrases: A dubious writing style emerging in science. Evidence of critical issues affecting established journals. arXiv. https://doi.org/10.48550/arXiv.2107.06751
Carrion, M. L. (2018). “You need to do your research”: Vaccines, contestable science, and maternal epistemology. Public Understanding of Science, 27(3), 310–324. https://doi.org/10.1177/0963662517728024
Chinn, S., & Hasell, A. (2023). Support for “doing your own research” is associated with COVID-19 misperceptions and scientific mistrust. Harvard Kennedy School (HSK) Misinformation Review, 4(3). https://doi.org/10.37016/mr-2020-117
Cholewiak, S. A., Ipeirotis, P., Silva, V., & Kannawadi, A. (2023). SCHOLARLY: Simple access to Google Scholar authors and citation using Python (1.5.0) [Computer software]. https://doi.org/10.5281/zenodo.5764801
Dadkhah, M., Lagzian, M., & Borchardt, G. (2017). Questionable papers in citation databases as an issue for literature review. Journal of Cell Communication and Signaling, 11(2), 181–185. https://doi.org/10.1007/s12079-016-0370-6
Dadkhah, M., Oermann, M. H., Hegedüs, M., Raman, R., & Dávid, L. D. (2023). Detection of fake papers in the era of artificial intelligence. Diagnosis, 10(4), 390–397. https://doi.org/10.1515/dx-2023-0090
Dunlap, R. E., & Brulle, R. J. (2020). Sources and amplifiers of climate change denial. In D. C. Holmes & L. M. Richardson (Eds.), Research handbook on communicating climate change (pp. 49–61). Edward Elgar Publishing. https://doi.org/10.4337/9781789900408.00013
Fares, M., Kutuzov, A., Oepen, S., & Velldal, E. (2017). Word vectors, reuse, and replicability: Towards a community repository of large-text resources. In J. Tiedemann & N. Tahmasebi (Eds.), Proceedings of the 21st Nordic Conference on Computational Linguistics (pp. 271–276). Association for Computational Linguistics. https://aclanthology.org/W17-0237
Google Scholar Help. (n.d.). Inclusion guidelines for webmasters. https://scholar.google.com/intl/en/scholar/inclusion.html
Gusenbauer, M., & Haddaway, N. R. (2020). Which academic search systems are suitable for systematic reviews or meta-analyses? Evaluating retrieval qualities of Google Scholar, PubMed, and 26 other resources. Research Synthesis Methods, 11(2), 181–217. https://doi.org/10.1002/jrsm.1378
Haider, J., & Åström, F. (2017). Dimensions of trust in scholarly communication: Problematizing peer review in the aftermath of John Bohannon’s “Sting” in science. Journal of the Association for Information Science and Technology, 68(2), 450–467. https://doi.org/10.1002/asi.23669
Huang, J., & Tan, M. (2023). The role of ChatGPT in scientific communication: Writing better scientific review articles. American Journal of Cancer Research, 13(4), 1148–1154. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10164801/
Jones, N. (2024). How journals are fighting back against a wave of questionable images. Nature, 626(8000), 697–698. https://doi.org/10.1038/d41586-024-00372-6
Kitamura, F. C. (2023). ChatGPT is shaping the future of medical writing but still requires human judgment. Radiology, 307(2), e230171. https://doi.org/10.1148/radiol.230171
Littell, J. H., Abel, K. M., Biggs, M. A., Blum, R. W., Foster, D. G., Haddad, L. B., Major, B., Munk-Olsen, T., Polis, C. B., Robinson, G. E., Rocca, C. H., Russo, N. F., Steinberg, J. R., Stewart, D. E., Stotland, N. L., Upadhyay, U. D., & Ditzhuijzen, J. van. (2024). Correcting the scientific record on abortion and mental health outcomes. BMJ, 384, e076518. https://doi.org/10.1136/bmj-2023-076518
Lund, B. D., Wang, T., Mannuru, N. R., Nie, B., Shimray, S., & Wang, Z. (2023). ChatGPT and a new academic reality: Artificial Intelligence-written research papers and the ethics of the large language models in scholarly publishing. Journal of the Association for Information Science and Technology, 74(5), 570–581. https://doi.org/10.1002/asi.24750
Martín-Martín, A., Orduna-Malea, E., Ayllón, J. M., & Delgado López-Cózar, E. (2016). Back to the past: On the shoulders of an academic search engine giant. Scientometrics, 107, 1477–1487. https://doi.org/10.1007/s11192-016-1917-2
Martín-Martín, A., Thelwall, M., Orduna-Malea, E., & Delgado López-Cózar, E. (2021). Google Scholar, Microsoft Academic, Scopus, Dimensions, Web of Science, and OpenCitations’ COCI: A multidisciplinary comparison of coverage via citations. Scientometrics, 126(1), 871–906. https://doi.org/10.1007/s11192-020-03690-4
Simon, F. M., Altay, S., & Mercier, H. (2023). Misinformation reloaded? Fears about the impact of generative AI on misinformation are overblown. Harvard Kennedy School (HKS) Misinformation Review, 4(5). https://doi.org/10.37016/mr-2020-127
Skeppstedt, M., Ahltorp, M., Kucher, K., & Lindström, M. (2024). From word clouds to Word Rain: Revisiting the classic word cloud to visualize climate change texts. Information Visualization, 23(3), 217–238. https://doi.org/10.1177/14738716241236188
Stokel-Walker, C. (2024, May 1.). AI Chatbots Have Thoroughly Infiltrated Scientific Publishing. Scientific American. https://www.scientificamerican.com/article/chatbots-have-thoroughly-infiltrated-scientific-publishing/
Subbaraman, N. (2024, May 14). Flood of fake science forces multiple journal closures: Wiley to shutter 19 more journals, some tainted by fraud. The Wall Street Journal. https://www.wsj.com/science/academic-studies-research-paper-mills-journals-publishing-f5a3d4bc
Thorp, H. H. (2023). ChatGPT is fun, but not an author. Science, 379(6630), 313–313. https://doi.org/10.1126/science.adg7879
Tripodi, F. B., Garcia, L. C., & Marwick, A. E. (2023). ‘Do your own research’: Affordance activation and disinformation spread. Information, Communication & Society, 27(6), 1212–1228. https://doi.org/10.1080/1369118X.2023.2245869
This research has been supported by Mistra, the Swedish Foundation for Strategic Environmental Research, through the research program Mistra Environmental Communication (Haider, Ekström, Rödl) and the Marcus and Amalia Wallenberg Foundation [2020.0004] (Söderström).
The research described in this article was carried out under Swedish legislation. According to the relevant EU and Swedish legislation (2003:460) on the ethical review of research involving humans (“Ethical Review Act”), the research reported on here is not subject to authorization by the Swedish Ethical Review Authority (“etikprövningsmyndigheten”) (SRC, 2017).
This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided that the original author and source are properly credited.
All data needed to replicate this study are available at the Harvard Dataverse: https://doi.org/10.7910/DVN/WUVD8X
The authors wish to thank two anonymous reviewers for their valuable comments on the article manuscript as well as the editorial group of Harvard Kennedy School (HKS) Misinformation Review for their thoughtful feedback and input.
The algorithmic knowledge gap within and between countries: Implications for combatting misinformation
While understanding how social media algorithms operate is essential to protect oneself from misinformation, such understanding is often unevenly distributed. This study explores the algorithmic knowledge gap both within and between countries, using national surveys in the United States (N = 1,415), the United Kingdom (N = 1,435), South Korea (N = 1,798), and Mexico (N = 784).
Keep Reading
Stochastic lies: How LLM-powered chatbots deal with Russian disinformation about the war in Ukraine
Research on digital misinformation has turned its attention to large language models (LLMs) and their handling of sensitive political topics. Through an AI audit, we analyze how three LLM-powered chatbots (Perplexity, Google Bard, and Bing Chat) generate content in response to the prompts linked to common Russian disinformation narratives about the war in Ukraine.
Keep Reading
...
Read the original on misinforeview.hks.harvard.edu »
AI from The Basement: My latest side project, a dedicated LLM server powered by 8x RTX 3090 Graphic Cards, boasting a total of 192GB of VRAM. I built this with running Meta’s Llamma-3.1 405B in mind.
This blogpost was originally posted on my LinkedIn profile in July 2024.
Backstory: Sometime in March I found myself struggling to keep up with the mere 48GB of VRAM I had been relying on for almost a year in my LLMs experimentations. So, in a geeky-yet-stylish way, I decided to spend my money to build this thing of beauty. Questions swirled: Which CPU/Platform to buy? Does memory speed really matter? And why the more PCIe Lanes we have the better? Why 2^n number of GPUs matter in multi-GPU node setup (Tensor Parallelism, anyone?) How many GPUs, and how can I get all the VRAM in the world? Why are Nvidia cards so expensive and why didn’t I invest in their stock earlier? What inference engine to use (hint: it’s not just llama.cpp and not always the most well-documented option)?
After so many hours of research, I decided on the following platform:
* Asrock Rack ROMED8-2T motherboard with 7x PCIe 4.0x16 slots and 128 lanes of PCIe
* A mere trio of 1600-watt power supply units to keep everything running smoothly
* 8x RTX 3090 GPUs with 4x NVLinks, enabling a blistering 112GB/s data transfer rate between each pair
Now that I kinda have everything in order, I’m working on a series of blog posts that will cover the entire journey, from building this behemoth to avoiding costly pitfalls. Topics will include:
* The challenges of assembling this system: from drilling holes in metal frames and adding 30amp 240volt breakers, to bending CPU socket pins (don’t try this at home, kids!).
* Why PCIe Risers suck and the importance of using SAS Device Adapters, Redrivers, and Retimers for error-free PCIe connections.
* NVLink speeds, PCIe lanes bandwidth and VRAM transfer speeds, and Nvidia’s decision to block P2P native PCIe bandwidth on the software level.
* Benchmarking inference engines like TensorRT-LLM, vLLM, and Aphrodite Engine, all of which support Tensor Parallelism.
* Training and fine-tuning your own LLM.
P. S. I’m sitting here staring at those GPUs, and I just can’t help but think how wild tech progress has been. I remember being so excited to get a 60GB HDD back in 2004. I mean, all the movies and games I could store?! Fast forward 20 years, and now I’ve got more than triple that storage capacity in just one machine’s graphic cards… It makes me think, what will we be doing in another 20 years?!
Anyway, that’s why I’m doing this project. I wanna help create some of the cool stuff that’ll be around in the future. And who knows, maybe someone will look back on my work and be like “haha, remember when we thought 192GB of VRAM was a lot?”
...
Read the original on ahmadosman.com »
Sometimes I just want to put pixels on a screen. I don’t want to think about SDL this or OpenGL that—I just want to draw my pixel buffer and be done.
fenster, a tiny 2D canvas library by Serge Zaitsev, does just that. It’s a tiny drop-in header-only C/C++ file that weighs no more than 400 LOC of pretty readable code. It works with WinAPI, Cocoa, and X11. And it handles keyboard and mouse input, too!
Sometimes I want to do just a little more than draw pixels—maybe have a menu, some buttons, render text—and I don’t want to completely DIY but I still don’t want to think about SDL.
Fortunately, microui by rxi exists and handles the translation from GUI elements into a simple retargetable drawing bytecode. It’s similarly a small, drop-in library, weighing only 1500 LOC.
Unfortunately, the demo program uses SDL as a backend for the bytecode. I’d been meaning to see if I could instead use fenster but understanding what a “quad” was or what “glScissor” did seemed intimidating. The project went nowhere.
Then, as usual, Kartik and I had a small argument and that resulted in us creating the fenster backend for microui! I sent him a skeleton to show what I wanted to do and he did most of the heavy lifting for the OpenGL-like parts.
The result is a less than 250 LOC file that binds microui to fenster. It’s inspired by the SDL renderer demo, but with a couple of added functions to abstract away keys and mouse buttons. It’s hacky and there’s some stuff we still don’t understand, but it works! And by “works” I mean draws the expected demo windows, handles mouse hover and click, and handles keyboard input.
* How to determine when to render from the texture and when from the provided
drawing command’s color
* Mod keys like so that, for example, + renders
Check it out here. It’s designed to all be dropped directly into your project.
This blog is open source.
See an error? Go ahead and
propose a change.
...
Read the original on bernsteinbear.com »
He built the glider, incidentally, with a gift of $5 sent to him by an American Civil War veteran after a school essay he’d written about Robert E. Lee was published in the local paper. The war, after all, had ended only 44 years earlier.
In 1946, by which time he’d become a notable writer of science fiction, he published a story called ‘A Logic named Joe’, which described a global computer network with servers and terminals, that starts giving people the information that it thinks they ought to know as opposed to waiting for them to search for it - the Singularity, if you like, or maybe just Alexa. He also, as I recall, predicted reality TV somewhere.
And yet, despite predicting half of our world, as a father in the 1950s he could not imagine why his daughter - my mother - wanted to work.
This isn’t an uncommon observation - plenty of people have pointed out that vintage scifi is full of rocketships but all the pilots are men. 1950s scifi shows 1950s society, but with robots. Meanwhile, the interstellar liners have paper tickets, that you queue up to buy. With fundamental technology change, we don’t so much get our predictions wrong as make predictions about the wrong things. (And, of course, we now have neither trolleys nor personal gliders.)
I was reminded of this photo recently when I came across a RAND ‘long-range forecasting’ study, from 1964. The authors polled a range of experts on what the key developments in coming decades would be and when they’d happen. Fields addressed included space flight and medicine, but the most interesting in this context is what was then called ‘automation’ (the past tended to describe as ‘automatic’ what we would now call ‘computers’). The double-page spread below shows the conclusions (click to enlarge).
...
Read the original on www.ben-evans.com »
Singletons looking to shack up with their soulmates online have relied on two key routes in the past decade or so: take your chance on dating apps, or befriend as many mutuals as possible on social media, in the hope that you find the one.
But some have found a third way, using services such as Goodreads and Strava to meet partners with whom they hope to spend the rest of their lives. Those couples proved to be trendsetters. So-called hobby apps — built around activites such as running, reading or movie-going — are having a moment, and not just for love.
It’s all part of a broader movement as people grow tired of the “digital town square” offered on Twitter/X and other social media platforms. At a time when many are abandoning Elon Musk’s social network over his attitude to “free speech” (which some see as “amplifying hate”), competing apps such as Bluesky and Threads are having a resurgence in users.
Whereas some users are switching to Twitter replicas, others are seeking refuge in apps that promise to connect them to people with whom they have common interests. Running app Strava has seen user numbers grow 20% in a year, according to digital market intelligence firm Sensor Tower. That success has led it to add a messaging tool for users to keep in touch, alongside documenting their workouts. Knitting social network Ravelry, which is accessed through a number of third-party apps, has more than 9 million users. Goodreads has clocked up more than 150 million members.
Letterboxd, a film completist’s dream app, where you can tick off the latest movies you’ve seen, and review and rate them, alongside other cinephiles and the occasional famous actor or director, has gone from having 1.8 million users worldwide in March 2020 to more than 14 million users this summer. The app has grown its monthly active userbase 55% in a year, according to Sensor Tower.
“We really work hard on the tone and voice of everything we do, from community policy through to editorial through to our social, to guide folks in terms of how we want them to be around Letterboxd,” says Gemma Gracewood, the app’s editor-in-chief. “We talk about movies.”
And that’s refreshing in a world where politics and culture wars are being pushed at us through algorithms. “Social media users have been turning towards niche apps and spaces for a long time,” says Jess Maddox, assistant professor in digital media at the University of Alabama. “Paradoxically, as major platforms such as Twitter/X, YouTube, TikTok and Instagram push more algorithmically curated feeds, users may be less exposed to the content they want to see.”
The cosy nature of hobby apps, and the way they’re set up to share passions and pastimes, means they’re an altogether gentler place than the rough-and-tumble racism you can encounter on X with an errant tap. “It’s a way for people to connect via common interests,” says Dr Carolina Are, a social media researcher at the Centre for Digital Citizens at Northumbria University. It all means that the apps can spend less time, effort and money on content moderation — assuming that civility will be supreme — and instead focus on making the overall experience better.
“The thing about Letterboxd is there isn’t a ‘central town square’ like there is on X; it’s a very single-channel conversation,” says Gracewood. Comments happen in-line — similar to those on the Guardian and Observer websites — meaning that it’s less possible to performatively repost content into a main feed in order to encourage a pile-on. Similar situations exist on platforms such as Goodreads and Strava, where it’s possible to communicate with and message others, but not to publicly shame them easily.
Because hobby apps are nicer places to exist, people spend more time on them — and they can eventually turn into services that are more than advertised. That includes finding like-minded people with whom you’d want to spend your time romantically.
One reason that people may be starting to find love on apps not explicitly designed for that purpose is because the expectations are lower — and as such, the atmosphere is less sexually charged. “Dating apps seem like a dating supermarket, and something you have to do if you want to have some kind of connection,” says Are.
She points out that while dating apps are trying to shed their shallow reputation as places to hook up, they still lead with giant pictures of users to gauge compatibility. “A lot of people are becoming quite disillusioned with the fact you’re judged on looks,” she says. “In general, there is a bit of a disillusionment with platform-facilitated dating culture, because it seems very impersonal. It’s all facilitated by an algorithm. And it seems not to serve people very well.”
Hobby apps’ gain is dating apps’ loss, based on recent financial figures from Match Group, the company that operates the best-known dating services, including Tinder and Hinge. From an October 2021 peak of more than $175 a share, Match is now trading at nearer $36 a share. The firm announced job cuts of 6% in July due to dwindling paying users.
But the rot isn’t limited to the big beasts in the game. An analysis of the top 200 dating and social connection apps by Deutsche Bank — entitled Dating: The Dating Debate — Have We Hit Saturation Levels? — suggests that global downloads have plateaued.
It also helps that hobby apps feel like a more cohesive, kinder community. That’s not just because the people are kinder: Letterboxd has a set of moderators who are tasked with taking a “zero tolerance” approach to overt or coded hate speech, racism, homophobia, white supremacy, transphobia or any other marginalising attitudes.
Letterboxd has fewer than 10 staff moderating content, says Gracewood, and generally they don’t need to intervene often. “I can’t speak to whether we’ve benefited from cultural and mission shifts at other social media platforms, but I can say that from day one, we have always been very, very concerned with what creating a community online looks like, and how to keep it feeling free, good and nice.”
Whether that light-touch approach compared with social media apps — TikTok employs 40,000 content moderators worldwide, while Meta reportedly has 15,000 — will last is yet to be seen. “It seems like every app is born, isn’t moderated, then something bad happens and it gets heavily moderated,” says Are. “So maybe they [hobby apps] will have that trajectory as well.”
Chris Stokel-Walker is the author of TikTok Boom: China’s Dynamite App and the Superpower Race for Social Media (Canbury Press, £9.99). To support the Guardian and Observer, order your copy at guardianbookshop.com. Delivery charges may apply
...
Read the original on www.theguardian.com »
When I peer into the far reaches of science fictional imagination, way out beyond the easy extrapolations and consensus futures, beyond the Blade Runners and the Star Treks, the name that looms largest is Iain M. Banks.
For those unfamiliar with his work in this genre, I’ll tell you a little bit about the Culture novels and recommend a reading approach. Then, for Banks beginners and devoted Culturephiles alike, I’ll explain why his future means so much to me.
What is the Culture? A civilization. An agreement. The subject of a collection of books, written across decades, which offer clues and suggestions, glances and reflections. A big part of the fun of reading those books is assembling your own mosaic. Here’s mine:
The Culture is a spacefaring, freewheeling admixture of anarchism and socialism. In most ways, it promises its citizens radical, breathtaking freedom … but in a few other ways, it requires their submission — to superhuman systems of planning and manufacture, the Culture’s ineffable Minds.
The Culture is a utopia: a future you might actually want to live in. It offers a coherent political vision. This isn’t subtle or allegorical; on the page, citizens of the Culture very frequently articulate and defend their values. (Their enthusiasm for their own politics is considered annoying by most other civilizations.)
Coherent political vision doesn’t require a lot, just some sense of this is what we ought to do, yet it is absent from plenty of science fiction that dwells only in the realm of the cautionary tale.
I don’t have much patience left for that genre. I mean … we have been, at this point, amply cautioned.
Vision, on the other hand: I can’t get enough.
The Culture novels aren’t connected by an overarching plot, and there is no canonical reading order. For all my appreciation: I have not even read all of them! If you search online, you’ll find plenty of proposed approaches.
Here is mine, which is unorthodox; call it a recipe for enjoying the Culture. It proceeds in three stages:
I very strongly believe new readers ought to start with Player of Games. It is a captivating novel in its own right, and its introduction to the Culture is smooth, almost stealthy. It’s also the book I started with, and obviously It Worked for Me, so I can’t help but recommend the same on-ramp.
For your second foray, you can choose basically at random. I like Matter and Surface Detail. I do not like Consider Phlebas.
Here is the unorthodox part: I don’t think it’s necessary, or even desirable, to have read more than a couple of Culture novels before turning to A Few Notes on the Culture, the post from Iain M. Banks that just … lays it all out there.
A Few Notes on the Culture is, for me, THE thrilling Culture document. It helps that it’s this odd sort of web samizdat — you are always reading a mirrored copy on some random website. The original was posted to a Usenet newsgroup in 1994!
I should say, I don’t generally love “raw worldbuilding” of this kind — RPG sourcebook material. This document is a brilliant exception, because the ideas are so big, so fresh, and so confidently articulated; and of course because it’s Iain M. Banks behind them, his voice inimitable, wry and winning.
Why not simply begin with A Few Notes on the Culture, if it’s so great? Well, it IS raw worldbuilding, and even the best exemplar of that genre benefits from narrative context. Read it on its own, and it’s a wonky thought experiment. Read it after a couple of novels, and it’s a backstage pass.
You ought to meet a character or two — hear from a few of the rollicking Minds, learn their wonderful names — before you go behind the curtain.
There are, in science fiction, several close peers to Iain M. Banks, at least in terms of the scale of their storytelling. I think in particular of Olaf Stapledon, his Last and First Men, which gallops across millions of years; and of Cixin Liu, his series starting with The Three-Body Problem, which bumps up against the death of the universe. I like both of these authors, but/and their futures are cold and grim. You wouldn’t call either one utopia.
So, I suppose it’s not just the scale of Iain M. Banks’s stories that I want to praise, but their warmth. His megastructures overflow with appealing characters pursuing interesting projects. Their voices are ironic and funny.
There’s no utopia without irony and humor; this fact really narrows the field.
In my novel Mr. Penumbra’s 24-Hour Bookstore, the ambitious and brilliant Kat Potente asks:
Kat straightens her shoulders. “Okay, we’re going to play. To start, imagine the future. The good future. No nuclear bombs. Pretend you’re a science fiction writer.”
“Go further. What’s the good future after that?”
“Star Trek. Transporters. You can go anywhere.”
Kat shakes her head. “It’s really hard. And that’s, what, a thousand years? What comes after that? What could possibly come after that? Imagination runs out. But it makes sense, right? We probably just imagine things based on what we already know, and we run out of analogies in the thirty-first century.”
I’m trying hard to imagine an average day in the year 3012. I can’t even come up with a half-decent scene. Will people live in buildings? Will they wear clothes? My imagination is almost physically straining.
Fingers of thought are raking the space behind the cushions, looking for loose ideas, finding nothing.
I have often described imagination as a muscle — one that, like any other muscle, can be developed.
Steady exposure to Star Trek gives you a minor workout, for sure; the bump of an imaginative bicep. But there’s much further to go. You can read your way into some of it, and some of it, you have to dream up for yourself.
Even among elite athletes, there must be titans: Schwarzeneggers striding across the stage. (I conjure bodybuilders because I like the idea of these imaginative muscles BULGING.) Iain M. Banks, who died in 2013, way too young, was Mr. Universe. Here was a great writer, sure; but here was an imagination unmatched.
He simply pushed further and thought bigger.
For me, the Culture is the standard, so, for a long time, the challenge has been implicit: can you, Sloan, imagine on that scale? And not just technically, but humanely — with warmth and irony?
I don’t get to Culture scale in Moonbound, but the plan — and there is a plan — is to ratchet up book by book, so my notional series can culminate in a feat that takes seriously the scale of the universe, as we now understand it.
It’s easier to write the defeat than the victory, isn’t it? Easier to write the failure than the success. For some reason, the success seems like it might be … boring.
Iain M. Banks shows us the Culture harnessing matter and energy on incredible scales. He tells us that citizens of the Culture live for hundreds of years; that death is generally a choice. In these books, !
At first glance, this seems fatal to plot. Endless energy, immortality … aren’t these the GOALS of the story? Are we just talking about heaven here? Heaven: which ought to be occluded, unknowable, unsayable. Heaven: because it’s boring.
Turns out, no, it’s not boring at all. Plot gallops on, even at the outer limits of matter and energy. Even at the far reaches of freedom, the stories are only just beginning.
In the writerly reticence to dramatize abundance, I detect humility — it requires serious imaginative muscle, beyond what most folks are working with — and I detect also cowardice. What if my utopia isn’t good enough? What if I say, “this is gonna be so great”, and readers reply, “eh, doesn’t sound that great”?
Easier to conjure some tyrant machines, some fast-spreading plagues. Everybody can agree on those.
Plenty of readers might indeed read the Culture novels and say, “eh, doesn’t sound that great”. The point is, there is something here to inspect, and consider, and, sure, even reject. In these novels, Iain M. Banks hoists the imaginative burden. He twirls it in the air. His muscles bulge. It’s amazing to behold.
P. S. Culturephiles will note I’ve mentioned only obiquely the Culture’s greatest aesthetic bounty, the names of its ships. That’s because they are actually not funny or interesting until you step inside the magic circle of the books, and begin to intuit the rules of the game. Of course, after you’ve read one or two Culture novels, learning new names becomes a large and growing fraction of the fun!
...
Read the original on www.robinsloan.com »
Having your photo taken at the Appalachian Trail Conservancy (ATC) headquarters in Harpers Ferry, West Virginia, has become a standard ritual for those hikers intending on walking the entire A. T. One of the functions of the ATC, as the lead organization in managing and protecting the A.T., is to maintain the official 2,000-miler registry of all those who have completed the A.T. Therefore, having a photo taken here makes many hikers feel as though their hikes have gained official recognition.
Thanks to a generous grant from the Quimby Family Foundation and the work of dedicated volunteers of the A. T. Museum under the leadership of Terry Harley Wilson, both scanned versions of the earlier Polaroid photos and the more recent digital photos can now be viewed on-line from anywhere in the world. When the A.T. Museum opened in Pennsylvania’s Pine Grove Furnace State Park in 2010, digital versions of the photos became available there too. (read expanded text)
...
Read the original on athikerpictures.org »
To add this web app to your iOS home screen tap the share button and select "Add to the Home Screen".
10HN is also available as an iOS App
If you visit 10HN only rarely, check out the the best articles from the past week.
If you like 10HN please leave feedback and share
Visit pancik.com for more.