Skip to main content
We’ve updated our Terms of Service. A new AI Addendum clarifies how Stack Overflow utilizes AI interactions.
Filter by
Sorted by
Tagged with
0 votes
1 answer
34 views

Upsert! Operation Throws "A table can't contain duplicate column names" Error

I have a base table A and a result table B in DolphinDB. Table B was initially empty and is used to store calculated results based on table A. When trying to insert the calculated results into table B,...
RORO's user avatar
  • 1
0 votes
0 answers
112 views

vLLM + Ray multi-node tensor-parallel deployment completely blocked by pending placement groups and raylet heartbeat failures

Environment: Ray version: 2.x vLLM version: 0.9.2 Python version: 3.9 OS / Container base: Linux (CentOS-based UBI8 in Kubernetes) Cloud / Infrastructure: AWS based Kubernetes cluster (pods scheduled ...
NullUser's user avatar
3 votes
1 answer
132 views

In Apache Ignite the Replication mode and Partition mode does not work all together

I’m working with Apache Ignite 2.17.0. I load database tables into Ignite caches and run SQL queries using the SQLFieldsQuery API. Recently, I modified the cache configuration for some tables to use ...
kushal Baldev's user avatar
1 vote
0 answers
32 views

Spark DSv2 Options vs Properties

I'm playing around with making a DSv2 data source, and I'm a bit confused about what the differences between the "options" and "properties" args passed to some of the TableProvider ...
William's user avatar
  • 141
0 votes
0 answers
62 views

Get two different nodes to access and distribute the same SQL table in Apache spark?

I have the following code to test. I created a table on worker 1. Then I tried to read the table on worker 2 and it got TABLE_OR_VIEW_NOT_FOUND. Worker 2 is in the some computer as Master. I ran the ...
Rick C. Ferreira's user avatar
3 votes
2 answers
230 views

How Ray async actors handle calls to sync methods

I'm working with Ray async actors and I want to understand exactly what happens—at a deep technical level—when a synchronous method is called on such an actor. I know that calling a synchronous method ...
hegash's user avatar
  • 893
0 votes
0 answers
50 views

How to best partition my data with a 32 core EMR instance and make sure I max out the parallelize feature?

I’m optimizing a PySpark pipeline that processes records with a heavily skewed categorical column (category). The data has: A few high-frequency categories (e.g., 90% of records fall into 2-3 ...
Bilal Jamil's user avatar
0 votes
0 answers
137 views

How to set up MS-MPI multi-machine communication between two Windows 11 systems?

I'm trying to set up a multi-machine communication environment using MS-MPI on two Windows 11 laptops, but I'm encountering some issues. Here are the details of my setup: Environment Details: ...
user29094781's user avatar
1 vote
1 answer
105 views

Distributed REST API Calls using SPARK with maintaining consistency

I have a Spark DataFrame created from a Delta table, with one column of type STRUCT(JSON). For each row in this DataFrame, I need to make a REST API call using the JSON payload in the column. ...
uds0128's user avatar
  • 53
0 votes
0 answers
18 views

MLP Speed-Up in PySpark fluctuates with more cores – possible cache memory issue?

enter image description here I have conducted experiments running the MLP (Multi-Layer Perceptron) algorithm on a PC cluster with Apache Spark, with configurations ranging from small data to large ...
Syahel Razaba's user avatar
0 votes
0 answers
321 views

PyTorch DDP Multi-Node Training: ncclInternalError: Internal check failed. Bootstrap : no socket interface found

I am trying to run a multi-node training job using PyTorch's DistributedDataParallel (DDP) following this guide. However, when I launch the job with torchrun, I encounter the following NCCL error on ...
yunjeong's user avatar
0 votes
1 answer
843 views

Clearing Cached Data on Databricks Cluster

The problem I am facing is that my "used" memory is only around 16GB, however the cached memory takes up so much space, that I am forced to use a compute with higher memory (64GB). So I ...
Manav Karthikeyan's user avatar
1 vote
0 answers
91 views

Segmentation Fault During Validation with MirroredStrategy on Multiple GPUs

I am training a model using TensorFlow 2.18.0 with the tf.distribute.MirroredStrategy across two GPUs. The training works fine on a single GPU, but when I try to run it on two GPUs, it ends with a ...
TGD's user avatar
  • 56
0 votes
1 answer
99 views

I want to use the distributed package in PyTorch for point-to-point communication between two ranks. but run error

def runTpoly(rank, size, pp, cs, pkArithmetics_evals, pkSelectors_evals, domain): init_process(rank, size) group2 = torch.distributed.new_group([1,2]) if rank == 0: device ...
wynne yin's user avatar
0 votes
0 answers
76 views

Vertex AI Reduction Server returning 500 Internal Error

I am looking to finetune a pre-trained deberta model on Vertex AI with pytorch. I'm attempting to run a distributed job, making use of the Vertex AI reduction server. I'm following this notebook: ...
purpleFudge's user avatar
0 votes
0 answers
55 views

How to develop multi-GPU modules in single-node single-GPU system in pytorch?

I'm developing a multi-GPU PyTorch application. Existing methods like scatter/gather in torch.distributed don't fulfill my requirements, therefore I need to develop forward/backprop steps which send ...
mirrortower's user avatar
1 vote
0 answers
31 views

Distributed training with Trainer and ConstantLengthDataset classes

I have a custom ConstantLengthDataset class: class ConstantLengthDataset(IterableDataset): def __init__( self, tokenizer, dataset, infinite=False, ...
имя's user avatar
  • 11
0 votes
2 answers
63 views

Interrupting busy worker process

I am trying to benchmark an algorithm I developed. To this end, I run the algorithm on several instances and measure time, memory, other numbers... For each instance, I create a new process; partly ...
Bubaya's user avatar
  • 893
0 votes
0 answers
78 views

How to properly handle multi-GPU execution failing on one GPU due to OOM

I'm working with multiple GPUs handling large amaounts of data. I want to create an out-of-memory (OOM) catch system that skips the current batch on all GPUs if any are out of memory. However, for ...
Zyzyx's user avatar
  • 534
0 votes
1 answer
36 views

How to maintain synchronization between distributed python processes?

I have a number of workstations that run long processes containing sequences like this: x = wait_while_current_is_set y = read_voltage z = z + y The workstations must maintain synchronization with a ...
david's user avatar
  • 2,697
0 votes
1 answer
475 views

Writing to a delta table spark 3.5.3 delta lake 3.2.0

It seems I'm unable to write using the delta format from my spark job, but I'm not sure what I'm missing. I'm using spark 3.5.3 and deltalake 3.2.0. My error: Exception in thread "main" org....
William's user avatar
  • 141
0 votes
1 answer
225 views

How to configure Ray cluster to utilize the Full Capacity of Databricks Cluster

I have a Databricks cluster configured with a minimum of 1 worker and a maximum of 4 workers, with auto-scaling enabled. What should my Ray configuration (setup_ray_cluster) be to fully utilize the ...
question.it's user avatar
  • 3,018
0 votes
0 answers
373 views

Cuda failure ‘named symbol not found’ when run on 4 L4 GPUs

Can someone help me with the following error. The code works fine on the 2 T4 GPUs. But fails when run on the 4 L4 GPUs. I am extending the Gemma 2B model for a multi-label multi-class classification ...
Rakesh Jarupula's user avatar
1 vote
0 answers
88 views

function to count maximum number of co-occurring entities for combination of 3.5 million ids?

I have been given this: x | y 1 | a,b,c,d,e 2 | a,b,c,d 3 | a,c,d ... I'd like this: 1,2 | 4 (a,b,c,d) 1,3 | 3 (a,c,d) 2,3 | 3 (a,c,d) I have x -> 3*10^6 such rows (3 million records) y could ...
Krithish Goli's user avatar
1 vote
0 answers
17 views

Why am I getting "TXN_REQUEST_IGNORED ERROR 10906" in GridDB due to an unknown event during cluster operations?

I’m using GridDB for a distributed database setup and recently encountered the following error while performing operations across nodes in the cluster: from griddb_python import StoreFactory, ...
Samar Mohamed's user avatar
0 votes
1 answer
70 views

Fault-tolerant queue-worker architecture in Kafka?

I am new to using queue-worker architectures and I'm interested in how to make it resilient to a worker failing. For example We have a pool of workers Alpha that put entries onto queue A Then the ...
Lubed Up Slug's user avatar
2 votes
2 answers
214 views

Client request failure in raft

Imagine a 3 node raft cluster. Each node is in sync has log [1,2,3] and entry 3 is committed by the leader. Now leader receives an entry 4 but fails to commit it because of unreliable network and ...
Dumb_Pegasus's user avatar
2 votes
0 answers
29 views

Why am I getting "SYNC_LOG_NOT_FOUND ERROR 20021" in GridDB during cluster synchronization?

I'm working with GridDB to manage a distributed cluster, and recently I’ve been encountering the following error during data synchronization across nodes in the cluster: 20021 SYNC_LOG_NOT_FOUND ERROR ...
Samar Mohamed's user avatar
0 votes
1 answer
133 views

Apache ignite compute broadcast

I'm trying apache ignite and must say ignite documentation is incomplete. Anyway, I've setup two node cluster using docker images 2.14.0-arm14 and exposed all Ports for both ignite containers, however ...
JUser's user avatar
  • 196
1 vote
0 answers
79 views

Dask erring on GridSearchCV and RandomizedSearchCV

I am trying to train an xgboost model using dask. I already have the data transformed andhave prepared my data as follows: X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, ...
joao pereira's user avatar
2 votes
0 answers
416 views

How to Implement Distributed AutoFAISS with PySpark for Large-Scale Vector Indexing?

I’m working on a project that involves creating a vector search index for a massive dataset consisting of 1.3 trillion tokens. I want to use AutoFAISS in a distributed environment to handle the scale ...
Cauder's user avatar
  • 2,759
1 vote
1 answer
56 views

run a function on different nodes of a slurm cluster for different parameters

How do I call the function inner from outer, such that each call to inner runs on a different node? That is, for ij = 1, it runs on node 1 using all of its 16 cores, for ij = 2, it runs on node 2 ...
evening silver fox's user avatar
0 votes
1 answer
100 views

Do I need Apache Spark to execute my Airflow DAG tasks?

I have a workflow with multiple DAGs. Every DAG has multiple tasks. These tasks are simple ETL tasks. It involves geo data in the form of kmls, csvs. An example task: We have meta data of road ...
ShariqHameed's user avatar
0 votes
1 answer
48 views

reliable protocol guarantee complete delivery no in order promise

The sender is sending N packets to receiver. I want a protocol or method that guarantees delivery, each packet is received at least once. It is ok if some packets are received more than once due to ...
Yufei Zheng's user avatar
0 votes
0 answers
35 views

How to securely conduct lottery-like draws with guaranteed randomness without auditing?

Is there an existing algorithm or method to conduct lottery-like draws that ensures secure and truly random results without the need for auditing? There are any lib to do this? I search on the web ...
aguiadouro's user avatar
0 votes
2 answers
534 views

Unable to run code on Multiple GPUs in PyTorch - Usage shows only 1 GPU is being utilized

I am training a Transformer Encoder-Decoder based model for Text summarization. The code works without any errors but uses only 1 GPU when checked with nvidia-smi. However, I want to run it on all the ...
Abid Meraj's user avatar
1 vote
2 answers
917 views

How to reliably implement fan out write pattern?

I'm trying to RELIABLY implement that pattern. For practical purposes, assume we have something similar to a twitter clone (in cassandra and nodejs). So, user A has 500k followers. When user A posts a ...
InglouriousBastard's user avatar
1 vote
1 answer
405 views

Using torchrun with AWS sagemaker estimator on multi-GPU node

I would like to run a training job ml.p4d.24xlarge machine on AWS SageMaker. I ran into a similar issue described here with significant slowdowns in training time. I understand now that I should run ...
probably45's user avatar
2 votes
0 answers
25 views

Why am I getting a "LM_WRITE_LOG_FAILED ERROR 80000" in GridDB when writing to the log file?

I'm using GridDB for managing a distributed database system and recently encountered the following error while trying to perform operations: 80000 LM_WRITE_LOG_FAILED ERROR Writing to log file failed. ...
omar esawy's user avatar
2 votes
0 answers
21 views

Why am I getting a "JC_CONTAINER_NOT_OPENED ERROR 145034" in GridDB when performing operations on a container?

I'm working with GridDB to manage a distributed database and recently encountered the following error while performing operations on a container: 145034 JC_CONTAINER_NOT_OPENED ERROR Status check of ...
omar esawy's user avatar
2 votes
0 answers
42 views

Why am I getting a "SYNC_CREATE_CONTEXT_FAILED ERROR 20037" during data synchronization in my GridDB cluster?

I'm working on a distributed system where I need to synchronize data across a cluster of nodes. However, I'm encountering an error during the synchronization process. The error message I get is: 20037 ...
omar esawy's user avatar
3 votes
1 answer
76 views

Async leader election in unrooted spanning tree declares multiple winners

I am trying to implement the algorithm described in the image, using MPI. It is part of a University project where we are building a distributed satellite to ground station communication system. I ...
Stelios Papamichail's user avatar
1 vote
0 answers
96 views

How to properly clean up non-serializable states associated with a Ray object?

Suppose I have a Ray actor that can create a Ray object that associates with some non-serializable states. In the following example, the non-serializable state is a temporary directory. class MyObject:...
Yang Bo's user avatar
  • 3,773
1 vote
1 answer
117 views

Error: invalid child spec in supervisor start_child function

In my erlang application i have a top level supervisor that monitors a cowboy server (gen_server): start_link() -> supervisor:start_link({local, ?SERVER}, ?MODULE, []). init([]) -> ...
salbh's user avatar
  • 71
2 votes
0 answers
99 views

How does XGBoost aggregate models being trained in a distributed fashion across n machines?

I am trying to understand how XGBoost distributed training works. The best explanation I've found so far is in this paper: https://ml-pai-learn.oss-cn-beijing.aliyuncs.com/%E6%9C%BA%E5%99%A8%E5%AD%A6%...
Altamash Rafiq's user avatar
0 votes
2 answers
114 views

Why does decreasing partition count prevent a StageFailure due to large size of serialized results?

I am trying to generate a randomly sorted version of a large-ish dataframe on databricks. My go-to code is to use .orderBy(rand()) on the dataframe. This, however, seems to trigger a SparkException ...
Felipe's user avatar
  • 11.9k
0 votes
1 answer
158 views

Micrometer & Prometheus with Java subprocesses that can't expose HTTP

I have a complex product that runs like this. A parent Java process which expose an HTTP service. The parent process starts worker subprocesses (new JVM) and manage the lifecycle of them. Worker ...
Joey Liu's user avatar
  • 510
0 votes
1 answer
109 views

Least Connection Load balancing using Grpc

least_connection.proto code Node overloaded -- starting load balancing process Traceback (most recent call last): File "D:\lab7p2\least connection\node2.py", line 73, in <module> node....
Yash Pahlani's user avatar
1 vote
1 answer
2k views

How to debug ValueError: `FlatParameter` requires uniform dtype but got torch.float32 and torch.bfloat16?

I'm trying to do Pytorch Lightning Fabric distributed FSDP training with Huggingface PEFT LORA fine tuning on LLAMA 2 but my code ends up failing with: `FlatParameter` requires uniform dtype but got ...
JobHunter69's user avatar
  • 2,366
0 votes
1 answer
49 views

Load pre-training parameters trained on a single GPU on multi GPUS on a single machine

I tried to load the pre-training parameters trained by a single GPU on a single machine with multiple GPUs, but errors such as Missing keys and Unexpected keys occurred. backbone_cfg = dict( ...
Mingshuai Zhao's user avatar

1
2 3 4 5
58