There is a large file that needs to be processed, with over 500 000 rows. Each row needs to be verified for various data and then all the results collated and reported into a single report. What is the most efficient way to do this? I tried calling the processing function with Job , where each function jobbed off would report results to a different node of a common global. But the jobbed functions are not updating their respective rows even though I am passing the global name and root node. What is the most efficient way to process this large file?

0 1
0 1.1K

I am working through trying to use ZAUTHENTICATE.mac and LDAP.mac to do Delegated sign on into Ensemble. In reading over the samples and the documentation, I am not clearly finding on how to set the Appropriate Role from the LDAP group I return. Can someone help explain this part to me? If I have a user sign on, and I return a "Group" from the Authentication, how do I get that to transform into the Role I need for Ensemble.

Thanks

Scott Roth

0 1
0 521

Hello Global masters,

I have some question about How can I connect C# to Cache Object without VisM?

I am always using VisM when I want to connect C# to Cache Object becouse I learn it from my co-worker .

but When I asked something in here about Cache Object and someone worte "don't use VisM because it is old thing" like that.

So I want to know about that

Thank you

Minsu.

0 1
0 333

Thanks for all replies in advance. In monitoring our audit log we are getting the error below. We have 4 different servers running CACHE 2014.1.1 and this is happening on all 4 of them. I have not been able to find what process is trying to logon but it is happening every 30 minutes. We have no user processes that are running so this has to be a CACHE process. Does anyone have an Ideal what could be running to cause this? What kind of trace can I run to determine what process is running it?

0 1
0 785

I have a lookup table and record batch Table I would like to do a count on the records stored in that batch by counting the number of records in that batch that have a certain key on the responseKey column. This column keys are stored in the look up table for comparison. So I would like to do a join sql pivot that will use my keys stored in the lookup as columns and count as values

so far I have managed to do this but this is not efficient I would like to fire that sql once not on every count

0 1
0 1.3K

Hello all,

I have an XSL question and I wanted to see if someone within a community can help me out.

To Summarize: I am working with XSL documents, single and multi-pages.
Inside my style sheet XML document, which is supporting my pages, everything on the page, including a headers, footers and the items are calculated
based on the portrait measurements.
Pages come up on the screen as portrait, but sometimes, depending on dynamic data are wide, they do not print properly in portrait.

0 1
0 398