To my surprise when you search for gleam read file in google, they are not much helpful information in the first page and no code example.
There are a post in Erlang Forums where the author of Gleam language pointed to a module that no longer exists in gleam_erlang pacakge, and a abandoned pacakge call gleam_file, and a couple pacakges like simplifile.
It turns out that Gleam has excellent FFI that if you are running it on BEAM (which is the default option unless you compile Gleam to javascript), for a simple case you just need to import the function from erlang, in just two lines of code.
Gleam language: how to find the min element in a list
Gleam language standard library has a
list.max
to find the maximum element in a list, but to my surprise it doesn’t provide a counterpart list.min function, in order to do that, you have to use compare function with order.negate
importgleam/listimportgleam/intimportgleam/orderpubfnmain(){letnumbers=[5,3,8,1,9,2]// Find the minimum value using list.max with order.negate
letminimum=list.max(numbers,with: fn(a,b){order.negate(int.compare(a,b))})// Print the result (will be Some(1))
io.debug(minimum)}
Another noteworthy aspect is that when a list contains multiple maximum values, list.max returns the last occurrence of the maximum value. This behavior contrasts significantly with Python’s list.max, which returns the first occurrence in such cases. I observed this discrepancy while comparing different implementations in both languages.
Transducers originated in Clojure, designed to tackle specific challenges in functional programming and data processing. If you’re working with large datasets, streaming data, or complex transformations, understanding transducers can significantly enhance the efficiency and composability of your code.
What Are Transducers?
At their core, transducers are composable functions that transform data. Unlike traditional functional programming techniques like map, filter, and reduce, which are tied to specific data structures, transducers abstract the transformation logic from the input and output, making them highly reusable and flexible.
Key Advantages of Transducers
1. Composability and Reusability
Transducers allow you to compose and reuse transformation logic across different contexts. By decoupling transformations from data structures, you can apply the same logic to lists, streams, channels, or any other sequential data structure. This makes your code more modular and adaptable.
2. Performance Optimization
One of the primary motivations for using transducers is to optimize data processing. Traditional approaches often involve creating intermediate collections, which can be costly in terms of performance, especially with large datasets. Transducers eliminate this overhead by performing all operations in a single pass, without generating intermediate results.
A Python example
importtimefromfunctoolsimportreduce# Traditional approachdeftraditional_approach(data):return[x*2forxindataif(x*2)%2==0]# Transducer approachdefmapping(f):deftransducer(reducer):defwrapped_reducer(acc,x):returnreducer(acc,f(x))returnwrapped_reducerreturntransducerdeffiltering(pred):deftransducer(reducer):defwrapped_reducer(acc,x):ifpred(x):returnreducer(acc,x)returnaccreturnwrapped_reducerreturntransducerdefcompose(t1,t2):defcomposed(reducer):returnt1(t2(reducer))returncomposeddeftransduce(data,initial,transducer,reducer):transformed_reducer=transducer(reducer)returnreduce(transformed_reducer,data,initial)data=range(1000000)# Measure traditional approachstart=time.time()traditional_result=traditional_approach(data)traditional_time=time.time()-start# Measure transducer approachxform=compose(mapping(lambdax:x*2),filtering(lambdax:x%2==0))defefficient_reducer(acc,x):acc.append(x)returnaccstart=time.time()transducer_result=transduce(data,[],xform,efficient_reducer)transducer_time=time.time()-start# Resultsprint(f"Traditional approach time: {traditional_time:.4f} seconds")print(f"Transducer approach time: {transducer_time:.4f} seconds")print(f"Traditional is faster by: {transducer_time/traditional_time:.2f}x")
however when executed the transducer version is much slower in Python
Traditional approach time: 0.0654 seconds
Transducer approach time: 0.1822 seconds
Traditional is faster by: 2.78x
Are Transducers Suitable for Python?
While transducers offer theoretical benefits in terms of composability and efficiency, Python might not be the best language for leveraging these advantages. Here’s why:
Python’s Function Call Overhead:
Python has a relatively high overhead for function calls. Since transducers rely heavily on higher-order functions, this overhead can negate the performance gains that transducers are designed to offer.
Optimized Built-in Functions:
Python’s built-in functions like map, filter, and list comprehensions are highly optimized in C. These built-ins often outperform custom transducer implementations, especially for common tasks.
Efficient Mutation with Lists:
Python’s lists are mutable, and appending to a list in a loop is highly efficient. The traditional method of using list comprehensions or filter and map is often faster and more straightforward than setting up a transducer pipeline.
When to Use Transducers
Transducers shine in functional programming languages that emphasize immutability and composability, such as Clojure or Gleam. In these languages, transducers can significantly reduce the overhead of creating intermediate collections and improve performance in complex data pipelines. They’re especially powerful when working with immutable data structures, where avoiding unnecessary copies is crucial for efficiency.
In contrast, Python’s strength lies in its mutable data structures and optimized built-in functions, which often make traditional approaches more performant. However, if you’re working in a functional programming environment where immutability is the norm, or if you need to maintain a consistent API across various data sources, transducers can be a valuable tool.
Conclusion
Transducers are a powerful tool in the right context, but Python’s inherent characteristics—such as function call overhead and optimized built-ins—mean that traditional approaches may be more efficient for typical data processing tasks. If you’re working in a language that deeply benefits from transducers, like Gleam, they can greatly enhance your code. In Python, however, it’s often best to use the language’s strengths, such as list comprehensions and optimized built-ins, for performance-critical applications.
The language is so straightforward that an experienced programmer can learn it in just a day or two. However, to truly appreciate its simplicity and constraints, one must have prior experience with complex programming languages and substantial coding practice.
focus on concrete problem
It is hard for less experienced developers to appreciate how rarely architecting for future requirements / applications turns out net-positive.
Racket provides layers of abstraction like rename-out to override operators like + - * / > < = (literally can be anything) when exporting module. This is great for building abstraction to teach concepts or to build another new language/DSL, but such flexibility often comes with maintainability costs and cognitive burden.
The Little Learner provides different implementations for tensors. Functions, and operators that appear identical have different meanings across different tensor implementations. Also operators are overridden, when reading the codebase, I often get confused by operators like <, sometimes it compare numbers, other times it compare scalars or tensors, Racket is a dynamic language, without type annotation, checking those functions and operations can be really frustrating and confusing.
While this uniform abstraction layer is beneficial for teaching machine learning concepts, it can be challenging when examining the actual code.
In contrast, Gleam shines with its simplicity and lack of hidden complexities. Everything must be explicitly stated, making the code clean and readable. Additionally, the compiler is smart enough to perform type inference, so you usually don’t need to add type notations for everything.
In [[Gleam]], records are compared by value (deep nested comparison), which can present challenges when using them as dictionary keys, unlike in some other functional languages.
Records are Compared by Value
It’s important to note that Gleam doesn’t have objects in the traditional sense. All data structures, including records, are compared by value. This means that two records with identical field values will be considered equal.
To make a record unique, an additional identifier field is necessary. This approach allows for distinguishing between records that might otherwise have the same content but represent different entities.
Ensuring Uniqueness
Simple Approach: UUID Field
One straightforward method to ensure record uniqueness is to add a UUID field. However, UUID strings can be memory-intensive and cpu-costly.
Improved Approach: Erlang Reference
A more efficient alternative is to use an [[erlang reference]] as a unique identifier for records.
Erlang references are unique identifiers created by the Erlang runtime system. They have several important properties:
Uniqueness: Each reference is guaranteed to be unique within an Erlang node (and even across connected nodes).
Lightweight: References are very memory-efficient.
Unguessable: They can’t be forged or guessed, which can be useful for security in some contexts.
Erlang-specific: They are native to the BEAM VM, so they work well with Gleam, which runs on this VM.
It’s important to note that:
Erlang references are not persistent across program runs. If you need to save and reload your records, you’ll need to implement a serialization strategy.
References are not garbage collected until the object they’re associated with is no longer referenced.
result.unwrap and result.or are both useful functions in Gleam for working with Result types, but they serve different purposes.
result.unwrap
result.unwrap is used to extract the value from a Result, providing a default value if the Result is an Error. It’s typically used when you want to proceed with a default value rather than propagating an error.
In this example, result.unwrap allows us to use a default value (“Guest”) when the user lookup fails, ensuring that we always have a name to greet.
result.or
result.or is used to provide an alternative Result when the first Result is an Error. It’s typically used when you have a fallback operation or value that you want to try if the primary operation fails.
Typical usage:
importgleam/resultpubfnget_config_from_file()->Result(Config,String){// Simulated file read
Error("File not found")}pubfnget_default_config()->Result(Config,String){// Return a default configuration
Ok(Config(..))}pubfnget_config()->Result(Config,String){get_config_from_file()|>result.or(get_default_config())}// Usage:
pubfnmain(){caseget_config(){Ok(config)->io.println("Config loaded")Error(err)->io.println("Failed to load config: "<>err)}}
In this example, result.or allows us to try loading the configuration from a file first, and if that fails, fall back to using a default configuration. The get_config function will only return an Error if both operations fail.
Key differences and when to use each:
Use result.unwrap when you want to extract a value from a Result and have a sensible default to use if it’s an Error. This effectively “throws away” the error information.
Use result.or when you want to try an alternative operation if the first one fails, while still preserving the Result type. This allows you to chain multiple fallback options.
result.unwrap returns the unwrapped value directly, while result.or returns another Result.
result.unwrap is often used at the “edges” of your program where you need to interface with code that doesn’t use Results, while result.or is more commonly used within the “core” logic where you’re still working with Results.
Both functions are valuable tools for error handling in Gleam, and understanding when to use each can lead to more robust and expressive code.
Rust
Same principle can be applied to Rust as the design is very alike.
(use-package!gleam-ts-mode:config;; setup formatter to be used by `SPC c f`(after!apheleia(setf(alist-get'gleam-ts-modeapheleia-mode-alist)'gleam)(setf(alist-get'gleamapheleia-formatters)'("gleam""format""--stdin"))))(after!treesit(add-to-list'auto-mode-alist'("\\.gleam$".gleam-ts-mode)))(after!gleam-ts-mode(unless(treesit-language-available-p'gleam);; compile the treesit grammar file the first time(gleam-ts-install-grammar)))
hack
If you, like me, use Treesitter grammar files from Nix, the tree-sitter subdirectory within the directory specified by user-emacs-directory is linked to Nix’s read-only filesystem, meaning gleam-ts-install-grammar is unable to install grammar files there.
Here’s how you can adjust treesit-extra-load-path and install the grammar file.
(after!gleam-ts-mode(setqtreesit-extra-load-path(list(expand-file-name"~/.local/tree-sitter/")))(unless(treesit-language-available-p'gleam);; hack: change `out-dir' when install language-grammar'(let((orig-treesit--install-language-grammar-1(symbol-function'treesit--install-language-grammar-1)))(cl-letf(((symbol-function'treesit--install-language-grammar-1)(lambda(out-dirlangurl)(funcallorig-treesit--install-language-grammar-1"~/.local/tree-sitter/"langurl))))(gleam-ts-install-grammar)))))