/dragonfly/sys/dev/drm/amd/amdgpu/ |
H A D | amdgpu_job.c | 54 if (!*job) in amdgpu_job_alloc() 62 (*job)->vm = vm; in amdgpu_job_alloc() 63 (*job)->ibs = (void *)&(*job)[1]; in amdgpu_job_alloc() 85 kfree(*job); in amdgpu_job_alloc_with_ib() 99 f = job->base.s_fence ? &job->base.s_fence->finished : job->fence; in amdgpu_job_free_resources() 114 kfree(job); in amdgpu_job_free_cb() 124 kfree(job); in amdgpu_job_free() 141 job->owner = owner; in amdgpu_job_submit() 159 r = amdgpu_ib_schedule(ring, job->num_ibs, job->ibs, NULL, fence); in amdgpu_job_submit_direct() 191 job); in amdgpu_job_dependency() [all …]
|
H A D | amdgpu_ib.c | 140 if (job) { in amdgpu_ib_schedule() 141 vm = job->vm; in amdgpu_ib_schedule() 142 fence_ctx = job->base.s_fence ? in amdgpu_ib_schedule() 154 if (vm && !job->vmid) { in amdgpu_ib_schedule() 180 if (job) { in amdgpu_ib_schedule() 188 if (job && ring->funcs->init_cond_exec) in amdgpu_ib_schedule() 205 status |= job->preamble_status; in amdgpu_ib_schedule() 220 amdgpu_ring_emit_ib(ring, ib, job ? job->vmid : 0, in amdgpu_ib_schedule() 237 if (job && job->uf_addr) { in amdgpu_ib_schedule() 238 amdgpu_ring_emit_fence(ring, job->uf_addr, job->uf_sequence, in amdgpu_ib_schedule() [all …]
|
H A D | amdgpu_job.h | 64 struct amdgpu_job **job, struct amdgpu_vm *vm); 66 struct amdgpu_job **job); 68 void amdgpu_job_free_resources(struct amdgpu_job *job); 69 void amdgpu_job_free(struct amdgpu_job *job); 70 int amdgpu_job_submit(struct amdgpu_job *job, struct drm_sched_entity *entity, 72 int amdgpu_job_submit_direct(struct amdgpu_job *job, struct amdgpu_ring *ring,
|
H A D | amdgpu_cs.c | 810 if (parser->job) in amdgpu_cs_parser_fini() 923 ib = &p->job->ibs[j]; in amdgpu_cs_ib_vm_chunk() 970 if (p->job->vm) { in amdgpu_cs_ib_vm_chunk() 1210 struct amdgpu_job *job; in amdgpu_cs_submit() local 1215 job = p->job; in amdgpu_cs_submit() 1216 p->job = NULL; in amdgpu_cs_submit() 1233 job->owner = p->filp; in amdgpu_cs_submit() 1240 amdgpu_job_free(job); in amdgpu_cs_submit() 1254 job->uf_sequence = seq; in amdgpu_cs_submit() 1275 job->base.s_fence = NULL; in amdgpu_cs_submit() [all …]
|
H A D | amdgpu_vm.c | 509 amdgpu_job_free(job); in amdgpu_vm_clear_bo() 725 if (job->vmid == 0) in amdgpu_vm_need_pipeline_sync() 801 trace_amdgpu_vm_flush(ring, job->vmid, job->vm_pd_addr); in amdgpu_vm_flush() 802 amdgpu_ring_emit_vm_flush(ring, job->vmid, job->vm_pd_addr); in amdgpu_vm_flush() 806 amdgpu_gmc_emit_pasid_mapping(ring, job->vmid, job->pasid); in amdgpu_vm_flush() 839 amdgpu_ring_emit_gds_switch(ring, job->vmid, job->gds_base, in amdgpu_vm_flush() 840 job->gds_size, job->gws_base, in amdgpu_vm_flush() 841 job->gws_size, job->oa_base, in amdgpu_vm_flush() 842 job->oa_size); in amdgpu_vm_flush() 1215 amdgpu_job_free(job); in amdgpu_vm_update_directories() [all …]
|
H A D | amdgpu_ids.c | 273 struct amdgpu_job *job, in amdgpu_vmid_grab_reserved() argument 318 job->vm_needs_flush = needs_flush; in amdgpu_vmid_grab_reserved() 338 struct amdgpu_job *job, in amdgpu_vmid_grab_used() argument 387 job->vm_needs_flush |= needs_flush; in amdgpu_vmid_grab_used() 408 struct amdgpu_job *job) in amdgpu_vmid_grab() argument 445 job->vm_needs_flush = true; in amdgpu_vmid_grab() 451 id->pd_gpu_addr = job->vm_pd_addr; in amdgpu_vmid_grab() 454 if (job->vm_needs_flush) { in amdgpu_vmid_grab() 458 job->vmid = id - id_mgr->ids; in amdgpu_vmid_grab() 459 job->pasid = vm->pasid; in amdgpu_vmid_grab() [all …]
|
H A D | amdgpu_vcn.c | 291 struct amdgpu_job *job; in amdgpu_vcn_dec_send_msg() local 300 ib = &job->ibs[0]; in amdgpu_vcn_dec_send_msg() 329 amdgpu_job_free(job); in amdgpu_vcn_dec_send_msg() 472 struct amdgpu_job *job; in amdgpu_vcn_enc_get_create_msg() local 482 ib = &job->ibs[0]; in amdgpu_vcn_enc_get_create_msg() 516 amdgpu_job_free(job); in amdgpu_vcn_enc_get_create_msg() 524 struct amdgpu_job *job; in amdgpu_vcn_enc_get_destroy_msg() local 534 ib = &job->ibs[0]; in amdgpu_vcn_enc_get_destroy_msg() 568 amdgpu_job_free(job); in amdgpu_vcn_enc_get_destroy_msg() 658 ib = &job->ibs[0]; in amdgpu_vcn_jpeg_set_reg() [all …]
|
H A D | amdgpu_vce.c | 435 struct amdgpu_job *job; in amdgpu_vce_get_create_msg() local 445 ib = &job->ibs[0]; in amdgpu_vce_get_create_msg() 486 r = amdgpu_job_submit_direct(job, ring, &f); in amdgpu_vce_get_create_msg() 496 amdgpu_job_free(job); in amdgpu_vce_get_create_msg() 514 struct amdgpu_job *job; in amdgpu_vce_get_destroy_msg() local 523 ib = &job->ibs[0]; in amdgpu_vce_get_destroy_msg() 547 r = amdgpu_job_submit_direct(job, ring, &f); in amdgpu_vce_get_destroy_msg() 560 amdgpu_job_free(job); in amdgpu_vce_get_destroy_msg() 710 struct amdgpu_ib *ib = &p->job->ibs[ib_idx]; in amdgpu_vce_ring_parse_cs() 721 p->job->vm = NULL; in amdgpu_vce_ring_parse_cs() [all …]
|
H A D | amdgpu_ttm.c | 1984 struct amdgpu_job *job; in amdgpu_map_buffer() local 2010 src_addr += job->ibs[0].gpu_addr; in amdgpu_map_buffer() 2017 amdgpu_ring_pad_ib(ring, &job->ibs[0]); in amdgpu_map_buffer() 2023 &job->ibs[0].ptr[num_dw]); in amdgpu_map_buffer() 2037 amdgpu_job_free(job); in amdgpu_map_buffer() 2048 struct amdgpu_job *job; in amdgpu_copy_buffer() local 2072 job->vm_needs_flush = vm_needs_flush; in amdgpu_copy_buffer() 2094 amdgpu_ring_pad_ib(ring, &job->ibs[0]); in amdgpu_copy_buffer() 2107 amdgpu_job_free(job); in amdgpu_copy_buffer() 2125 struct amdgpu_job *job; in amdgpu_fill_buffer() local [all …]
|
H A D | amdgpu_trace.h | 44 #define trace_amdgpu_vm_grab_id(vm, ring_idx, job) argument 54 #define trace_amdgpu_sched_run_job(job) argument
|
/dragonfly/contrib/bmake/ |
H A D | job.c | 502 for (job = job_table; job < job_table_end; job++) { in DumpJobs() 505 (int)(job - job_table), job->status, flags, job->pid); in DumpJobs() 602 for (job = job_table; job < job_table_end; job++) { in JobCondPassSig() 723 for (job = job_table; job < job_table_end; job++) { in JobFindPid() 1105 meta_job_error(job, job->node, in JobFinishDoneExitedError() 1667 for (job = job_table; job < job_table_end; job++) { in JobStart() 1674 memset(job, 0, sizeof *job); in JobStart() 1849 nRead = read(job->inPipe, &job->outBuf[job->curPos], in CollectOutput() 2592 for (job = job_table; job < job_table_end; job++) { in JobInterrupt() 2679 for (job = job_table; job < job_table_end; job++) { in Job_AbortAll() [all …]
|
H A D | trace.c | 84 Trace_Log(TrEvent event, Job *job) in Trace_Log() argument 104 if (job != NULL) { in Trace_Log() 107 Job_FlagsToString(job, flags, sizeof flags); in Trace_Log() 108 fprintf(trfile, " %s %d %s %x", job->node->name, in Trace_Log() 109 job->pid, flags, job->node->type); in Trace_Log()
|
H A D | meta.c | 654 if (job != NULL) { in meta_job_start() 655 pbm = &job->bm; in meta_job_start() 662 if (job == NULL) in meta_job_start() 686 pbm = &job->bm; in meta_job_child() 711 pbm = &job->bm; in meta_job_parent() 728 pbm = &job->bm; in meta_job_fd() 746 pbm = &job->bm; in meta_job_event() 764 pbm = &job->bm; in meta_job_error() 766 gn = job->node; in meta_job_error() 792 pbm = &job->bm; in meta_job_output() [all …]
|
/dragonfly/usr.sbin/yppush/ |
H A D | yppush_main.c | 111 struct jobs *job; in yppush_show_status() local 118 job = job->next; in yppush_show_status() 121 if (job == NULL) { in yppush_show_status() 127 if (job->polled) { in yppush_show_status() 135 job->tid); in yppush_show_status() 140 job->map, job->server, status == YPXFR_SUCC ? in yppush_show_status() 147 job->polled = 1; in yppush_show_status() 398 job->stat = 0; in yp_push() 399 job->tid = tid; in yp_push() 404 job->polled = 0; in yp_push() [all …]
|
/dragonfly/bin/sh/ |
H A D | jobs.c | 98 struct job { struct 146 static struct job *getcurjob(struct job *); 261 struct job *jp; in fgcmd() 286 struct job *jp; in bgcmd() 391 struct job *j; in showjob() 493 struct job *jp; in showjobs() 552 struct job *job; in waitcmd() local 572 waitcmdloop(struct job *job) in waitcmdloop() argument 575 struct job *jp; in waitcmdloop() 759 struct job * [all …]
|
H A D | jobs.h | 43 struct job; 56 struct job *makejob(union node *, int); 57 pid_t forkshell(struct job *, union node *, int); 58 pid_t vforkexecshell(struct job *, char **, char **, const char *, int, int []); 59 int waitforjob(struct job *, int *);
|
/dragonfly/sys/dev/drm/scheduler/ |
H A D | gpu_scheduler.c | 242 drm_sched_fence_finished(job->s_fence); in drm_sched_entity_kill_jobs_cb() 243 WARN_ON(job->s_fence->parent); in drm_sched_entity_kill_jobs_cb() 244 dma_fence_put(&job->s_fence->finished); in drm_sched_entity_kill_jobs_cb() 245 job->sched->ops->free_job(job); in drm_sched_entity_kill_jobs_cb() 317 struct drm_sched_job *job; in drm_sched_entity_fini() local 593 schedule_work(&job->finish_work); in drm_sched_job_finish_cb() 617 job->sched->ops->timedout_job(job); in drm_sched_job_timedout() 743 job->sched = sched; in drm_sched_job_init() 744 job->entity = entity; in drm_sched_job_init() 747 if (!job->s_fence) in drm_sched_job_init() [all …]
|
H A D | gpu_scheduler_trace.h | 36 #define trace_amd_sched_job(job) argument 37 #define trace_amd_sched_process_job(job) argument
|
/dragonfly/contrib/zstd/lib/compress/ |
H A D | zstdmt_compress.c | 689 if (job->cdict) { in ZSTDMT_compressionJob() 694 U64 const pledgedSrcSize = job->firstJob ? job->fullFrameSize : job->src.size; in ZSTDMT_compressionJob() 707 ZSTDMT_serialState_update(job->serial, cctx, rawSeqStore, job->src, job->jobID); in ZSTDMT_compressionJob() 726 assert(job->cSize == 0); in ZSTDMT_compressionJob() 755 ZSTDMT_serialState_ensureFinished(job->serial, job->jobID, job->cSize); in ZSTDMT_compressionJob() 756 if (job->prefix.size > 0) in ZSTDMT_compressionJob() 1305 assert(job->lastJob == 1); in ZSTDMT_writeLastEmptyBlock() 1309 job->dstBuff = ZSTDMT_getBuffer(job->bufPool); in ZSTDMT_writeLastEmptyBlock() 1315 job->src = kNullRange; in ZSTDMT_writeLastEmptyBlock() 1316 job->cSize = ZSTD_writeLastEmptyBlock(job->dstBuff.start, job->dstBuff.capacity); in ZSTDMT_writeLastEmptyBlock() [all …]
|
/dragonfly/usr.sbin/cron/cron/ |
H A D | job.c | 28 } job; typedef 31 static job *jhead = NULL, *jtail = NULL; 37 job *j; in job_add() 44 if ((j = (job*)malloc(sizeof(job))) == NULL) in job_add() 60 job *j, *jn; in job_runqueue()
|
/dragonfly/contrib/tcsh-6/nls/ja/ |
H A D | set1 | 61 59 カレントのjobはありません 62 60 前のjobはありません 67 65 %s中断したjobが残っています 68 66 %s止まったjobが残っています 138 136 jobが見つかりません
|
/dragonfly/usr.bin/dsynth/ |
H A D | bulk.c | 39 typedef struct job { struct 338 job_t *job = arg; in bulkthread() local 343 if (job->terminate) in bulkthread() 345 if (job->active == NULL) in bulkthread() 346 pthread_cond_wait(&job->cond, &BulkMutex); in bulkthread() 347 bulk = job->active; in bulkthread() 366 if ((bulk = BulkSubmit) != NULL && job->terminate == 0) { in bulkthread() 371 job->active = bulk; in bulkthread() 374 job->active = NULL; in bulkthread()
|
/dragonfly/usr.sbin/lpr/lpc/ |
H A D | cmds.c | 1204 doarg(char *job) in doarg() argument 1216 if ((cp = strchr(job, ':')) != NULL) { in doarg() 1217 machine = job; in doarg() 1219 job = cp; in doarg() 1226 if (isdigit(*job)) { in doarg() 1229 jobnum = jobnum * 10 + (*job++ - '0'); in doarg() 1230 while (isdigit(*job)); in doarg() 1237 if (*job && strcmp(job, cp) != 0) in doarg() 1261 if (line[0] != 'P' || strcmp(job, line+1) != 0) in doarg()
|
/dragonfly/contrib/zstd/lib/common/ |
H A D | pool.c | 86 { POOL_job const job = ctx->queue[ctx->queueHead]; in POOL_thread() local 94 job.function(job.opaque); in POOL_thread() 262 POOL_job const job = {function, opaque}; in POOL_add_internal() local 267 ctx->queue[ctx->queueTail] = job; in POOL_add_internal()
|
/dragonfly/usr.sbin/cron/doc/ |
H A D | Makefile.vixie | 93 env.c job.c user.c popen.c misc.c compat.c 96 misc.c job.c do_command.c env.c popen.c 98 CRON_OBJ = cron.o database.o user.o entry.o job.o do_command.o \
|