Class: GraphQL::Dataloader
- Inherits:
-
Object
- Object
- GraphQL::Dataloader
- Defined in:
- lib/graphql/dataloader.rb,
lib/graphql/dataloader/source.rb,
lib/graphql/dataloader/request.rb,
lib/graphql/dataloader/request_all.rb,
lib/graphql/dataloader/null_dataloader.rb,
lib/graphql/dataloader/async_dataloader.rb
Overview
This plugin supports Fiber-based concurrency, along with Source.
Direct Known Subclasses
Defined Under Namespace
Classes: AsyncDataloader, NullDataloader, Request, RequestAll, Source
Class Attribute Summary collapse
-
.default_fiber_limit ⇒ Object
Returns the value of attribute default_fiber_limit.
-
.default_nonblocking ⇒ Object
Returns the value of attribute default_nonblocking.
Instance Attribute Summary collapse
- #fiber_limit ⇒ Integer? readonly
Class Method Summary collapse
- .use(schema, nonblocking: nil, fiber_limit: nil) ⇒ Object
-
.with_dataloading(&block) ⇒ Object
Call the block with a Dataloader instance, then run all enqueued jobs and return the result of the block.
Instance Method Summary collapse
- #append_job(&job) ⇒ Object
-
#cleanup_fiber ⇒ Object
This method is called when Dataloader is finished using a fiber.
-
#clear_cache ⇒ void
Clear any already-loaded objects from Source caches.
-
#get_fiber_variables ⇒ Hash<Symbol, Object>
This is called before the fiber is spawned, from the parent context (i.e. from the thread or fiber that it is scheduled from).
-
#initialize(nonblocking: self.class.default_nonblocking, fiber_limit: self.class.default_fiber_limit) ⇒ Dataloader
constructor
A new instance of Dataloader.
- #nonblocking? ⇒ Boolean
- #run ⇒ Object
- #run_fiber(f) ⇒ Object
-
#run_isolated ⇒ Object
Use a self-contained queue for the work in the block.
-
#set_fiber_variables(vars) ⇒ void
Set up the fiber variables in a new fiber.
- #spawn_fiber ⇒ Object
-
#with(source_class, *batch_args, **batch_kwargs) ⇒ Object
truffle-ruby wasn't doing well with the implementation below.
-
#yield ⇒ void
Tell the dataloader that this fiber is waiting for data.
Constructor Details
#initialize(nonblocking: self.class.default_nonblocking, fiber_limit: self.class.default_fiber_limit) ⇒ Dataloader
Returns a new instance of Dataloader.
58 59 60 61 62 63 64 65 |
# File 'lib/graphql/dataloader.rb', line 58 def initialize(nonblocking: self.class.default_nonblocking, fiber_limit: self.class.default_fiber_limit) @source_cache = Hash.new { |h, k| h[k] = {} } @pending_jobs = [] if !nonblocking.nil? @nonblocking = nonblocking end @fiber_limit = fiber_limit end |
Class Attribute Details
.default_fiber_limit ⇒ Object
Returns the value of attribute default_fiber_limit.
27 28 29 |
# File 'lib/graphql/dataloader.rb', line 27 def default_fiber_limit @default_fiber_limit end |
.default_nonblocking ⇒ Object
Returns the value of attribute default_nonblocking.
27 28 29 |
# File 'lib/graphql/dataloader.rb', line 27 def default_nonblocking @default_nonblocking end |
Instance Attribute Details
#fiber_limit ⇒ Integer? (readonly)
68 69 70 |
# File 'lib/graphql/dataloader.rb', line 68 def fiber_limit @fiber_limit end |
Class Method Details
.use(schema, nonblocking: nil, fiber_limit: nil) ⇒ Object
30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 |
# File 'lib/graphql/dataloader.rb', line 30 def self.use(schema, nonblocking: nil, fiber_limit: nil) dataloader_class = if nonblocking warn("`nonblocking: true` is deprecated from `GraphQL::Dataloader`, please use `GraphQL::Dataloader::AsyncDataloader` instead. Docs: https://graphql-ruby.org/dataloader/async_dataloader.") Class.new(self) { self.default_nonblocking = true } else self end if fiber_limit dataloader_class = Class.new(dataloader_class) dataloader_class.default_fiber_limit = fiber_limit end schema.dataloader_class = dataloader_class end |
.with_dataloading(&block) ⇒ Object
Call the block with a Dataloader instance, then run all enqueued jobs and return the result of the block.
48 49 50 51 52 53 54 55 56 |
# File 'lib/graphql/dataloader.rb', line 48 def self.with_dataloading(&block) dataloader = self.new result = nil dataloader.append_job { result = block.call(dataloader) } dataloader.run result end |
Instance Method Details
#append_job(&job) ⇒ Object
138 139 140 141 142 143 |
# File 'lib/graphql/dataloader.rb', line 138 def append_job(&job) # Given a block, queue it up to be worked through when `#run` is called. # (If the dataloader is already running, than a Fiber will pick this up later.) @pending_jobs.push(job) nil end |
#cleanup_fiber ⇒ Object
This method is called when Dataloader is finished using a fiber. Use it to perform any cleanup, such as releasing database connections (if required manually)
99 100 |
# File 'lib/graphql/dataloader.rb', line 99 def cleanup_fiber end |
#clear_cache ⇒ void
This method returns an undefined value.
Clear any already-loaded objects from Source caches
147 148 149 150 151 152 |
# File 'lib/graphql/dataloader.rb', line 147 def clear_cache @source_cache.each do |_source_class, batched_sources| batched_sources.each_value(&:clear_cache) end nil end |
#get_fiber_variables ⇒ Hash<Symbol, Object>
This is called before the fiber is spawned, from the parent context (i.e. from the thread or fiber that it is scheduled from).
78 79 80 81 82 83 84 |
# File 'lib/graphql/dataloader.rb', line 78 def get_fiber_variables fiber_vars = {} Thread.current.keys.each do |fiber_var_key| fiber_vars[fiber_var_key] = Thread.current[fiber_var_key] end fiber_vars end |
#nonblocking? ⇒ Boolean
70 71 72 |
# File 'lib/graphql/dataloader.rb', line 70 def nonblocking? @nonblocking end |
#run ⇒ Object
186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226 227 228 229 230 231 232 233 234 235 |
# File 'lib/graphql/dataloader.rb', line 186 def run jobs_fiber_limit, total_fiber_limit = calculate_fiber_limit job_fibers = [] next_job_fibers = [] source_fibers = [] next_source_fibers = [] first_pass = true manager = spawn_fiber do while first_pass || !job_fibers.empty? first_pass = false while (f = (job_fibers.shift || (((next_job_fibers.size + job_fibers.size) < jobs_fiber_limit) && spawn_job_fiber))) if f.alive? finished = run_fiber(f) if !finished next_job_fibers << f end end end join_queues(job_fibers, next_job_fibers) while (!source_fibers.empty? || @source_cache.each_value.any? { |group_sources| group_sources.each_value.any?(&:pending?) }) while (f = source_fibers.shift || (((job_fibers.size + source_fibers.size + next_source_fibers.size + next_job_fibers.size) < total_fiber_limit) && spawn_source_fiber)) if f.alive? finished = run_fiber(f) if !finished next_source_fibers << f end end end join_queues(source_fibers, next_source_fibers) end end end run_fiber(manager) if manager.alive? raise "Invariant: Manager fiber didn't terminate properly." end if !job_fibers.empty? raise "Invariant: job fibers should have exited but #{job_fibers.size} remained" end if !source_fibers.empty? raise "Invariant: source fibers should have exited but #{source_fibers.size} remained" end rescue UncaughtThrowError => e throw e.tag, e.value end |
#run_fiber(f) ⇒ Object
237 238 239 |
# File 'lib/graphql/dataloader.rb', line 237 def run_fiber(f) f.resume end |
#run_isolated ⇒ Object
Use a self-contained queue for the work in the block.
155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 |
# File 'lib/graphql/dataloader.rb', line 155 def run_isolated prev_queue = @pending_jobs prev_pending_keys = {} @source_cache.each do |source_class, batched_sources| batched_sources.each do |batch_args, batched_source_instance| if batched_source_instance.pending? prev_pending_keys[batched_source_instance] = batched_source_instance.pending.dup batched_source_instance.pending.clear end end end @pending_jobs = [] res = nil # Make sure the block is inside a Fiber, so it can `Fiber.yield` append_job { res = yield } run res ensure @pending_jobs = prev_queue prev_pending_keys.each do |source_instance, pending| pending.each do |key, value| if !source_instance.results.key?(key) source_instance.pending[key] = value end end end end |
#set_fiber_variables(vars) ⇒ void
This method returns an undefined value.
Set up the fiber variables in a new fiber.
This is called within the fiber, right after it is spawned.
92 93 94 95 |
# File 'lib/graphql/dataloader.rb', line 92 def set_fiber_variables(vars) vars.each { |k, v| Thread.current[k] = v } nil end |
#spawn_fiber ⇒ Object
241 242 243 244 245 246 247 248 |
# File 'lib/graphql/dataloader.rb', line 241 def spawn_fiber fiber_vars = get_fiber_variables Fiber.new(blocking: !@nonblocking) { set_fiber_variables(fiber_vars) yield cleanup_fiber } end |
#with(source_class, *batch_args, **batch_kwargs) ⇒ Object
truffle-ruby wasn't doing well with the implementation below
109 110 111 112 113 114 115 116 |
# File 'lib/graphql/dataloader.rb', line 109 def with(source_class, *batch_args) batch_key = source_class.batch_key_for(*batch_args) @source_cache[source_class][batch_key] ||= begin source = source_class.new(*batch_args) source.setup(self) source end end |
#yield ⇒ void
This method returns an undefined value.
Tell the dataloader that this fiber is waiting for data.
Dataloader will resume the fiber after the requested data has been loaded (by another Fiber).
132 133 134 135 |
# File 'lib/graphql/dataloader.rb', line 132 def yield Fiber.yield nil end |